neural decoding of visual imagery during sleep t. horikawa et al. … · 2013-07-10 · doi:...

39
DOI: 10.1126/science.1234330 , 639 (2013); 340 Science et al. T. Horikawa Neural Decoding of Visual Imagery During Sleep This copy is for your personal, non-commercial use only. clicking here. colleagues, clients, or customers by , you can order high-quality copies for your If you wish to distribute this article to others here. following the guidelines can be obtained by Permission to republish or repurpose articles or portions of articles ): July 10, 2013 www.sciencemag.org (this information is current as of The following resources related to this article are available online at http://www.sciencemag.org/content/340/6132/639.full.html version of this article at: including high-resolution figures, can be found in the online Updated information and services, http://www.sciencemag.org/content/suppl/2013/04/03/science.1234330.DC1.html http://www.sciencemag.org/content/suppl/2013/04/04/science.1234330.DC2.html can be found at: Supporting Online Material http://www.sciencemag.org/content/340/6132/639.full.html#related found at: can be related to this article A list of selected additional articles on the Science Web sites http://www.sciencemag.org/content/340/6132/639.full.html#ref-list-1 , 10 of which can be accessed free: cites 34 articles This article http://www.sciencemag.org/cgi/collection/neuroscience Neuroscience subject collections: This article appears in the following registered trademark of AAAS. is a Science 2013 by the American Association for the Advancement of Science; all rights reserved. The title Copyright American Association for the Advancement of Science, 1200 New York Avenue NW, Washington, DC 20005. (print ISSN 0036-8075; online ISSN 1095-9203) is published weekly, except the last week in December, by the Science on July 10, 2013 www.sciencemag.org Downloaded from

Upload: others

Post on 27-May-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

DOI: 10.1126/science.1234330, 639 (2013);340 Science et al.T. Horikawa

Neural Decoding of Visual Imagery During Sleep

This copy is for your personal, non-commercial use only.

clicking here.colleagues, clients, or customers by , you can order high-quality copies for yourIf you wish to distribute this article to others

here.following the guidelines

can be obtained byPermission to republish or repurpose articles or portions of articles

): July 10, 2013 www.sciencemag.org (this information is current as of

The following resources related to this article are available online at

http://www.sciencemag.org/content/340/6132/639.full.htmlversion of this article at:

including high-resolution figures, can be found in the onlineUpdated information and services,

http://www.sciencemag.org/content/suppl/2013/04/03/science.1234330.DC1.html http://www.sciencemag.org/content/suppl/2013/04/04/science.1234330.DC2.html

can be found at: Supporting Online Material

http://www.sciencemag.org/content/340/6132/639.full.html#relatedfound at:

can berelated to this article A list of selected additional articles on the Science Web sites

http://www.sciencemag.org/content/340/6132/639.full.html#ref-list-1, 10 of which can be accessed free:cites 34 articlesThis article

http://www.sciencemag.org/cgi/collection/neuroscienceNeuroscience

subject collections:This article appears in the following

registered trademark of AAAS. is aScience2013 by the American Association for the Advancement of Science; all rights reserved. The title

CopyrightAmerican Association for the Advancement of Science, 1200 New York Avenue NW, Washington, DC 20005. (print ISSN 0036-8075; online ISSN 1095-9203) is published weekly, except the last week in December, by theScience

on

July

10,

201

3w

ww

.sci

ence

mag

.org

Dow

nloa

ded

from

Page 2: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

antigen-presenting cells or cytokines. In eithercase, the current observations exclude models inwhich each naïve T cell yields progeny with thesame distribution of cells of either short- or long-term potential. Thus, although our data do notevaluate the role of asymmetric division as amechanism to generate daughter cells with dif-ferent fates, they do show that asymmetric divi-sion by itself cannot explain the disparity betweenindividual T cell families that is experimentallyobserved. Rather, a strong variation between fam-ilies in the expansion of proximal and distaldaughter cells needs to be invoked to arrive at theheterogeneity observed here. Lastly, the observa-tion of strong heterogeneity at the single-cell lev-el indicates that T cell responses are made up of“averages,” a behavior reminiscent of recent analy-ses of stem-cell renewal (19). Thus, although thedifferentiation and expansion of the combinedT cell population follow a uniform course, thefate of individual naïve T cells is unpredictable.

References and Notes1. J. J. Obar, K. M. Khanna, L. Lefrançois, Immunity 28, 859

(2008).2. J. N. Blattman et al., J. Exp. Med. 195, 657 (2002).3. K. Murali-Krishna et al., Immunity 8, 177 (1998).4. E. A. Butz, M. J. Bevan, Immunity 8, 167 (1998).5. M. A. Williams, M. J. Bevan, Annu. Rev. Immunol. 25,

171 (2007).6. S. Sarkar et al., J. Exp. Med. 205, 625 (2008).7. J. J. Obar, L. Lefrançois, Int. Immunol. 22, 619 (2010).8. C. Stemberger et al., Immunity 27, 985 (2007).9. C. Gerlach et al., J. Exp. Med. 207, 1235 (2010).10. J. T. Chang et al., Science 315, 1687 (2007).11. J. W. J. van Heijst et al., Science 325, 1265

(2009).12. T. N. M. Schumacher, C. Gerlach, J. W. J. van Heijst,

Nat. Rev. Immunol. 10, 621 (2010).13. K. Schepers et al., J. Exp. Med. 205, 2309 (2008).14. Materials and methods are available as supplementary

materials on Science Online.15. W. N. D’Souza, S. M. Hedrick, J. Immunol. 177, 777

(2006).16. T. E. Schlub et al., Eur. J. Immunol. 39, 67 (2009).17. V. R. Buchholz et al., Science 340, 630 (2013).18. K. R. Duffy et al., Science 335, 338 (2012).19. B. D. Simons, H. Clevers, Cell 145, 851 (2011).

Acknowledgments: The authors thank A. Pfauth andF. van Diepen for cell sorting; E. Borg for technical assistance;D. Zehn for provision of LM-OVA (T4); S. Harkema for helpfuldiscussions; and G. Bendle, M. C. Wolkers, and E. A. Mosemanfor critically reading the manuscript. The data presented inthis manuscript are tabulated in the main paper and in thesupplementary materials. The authors declare no conflict ofinterest. This research was funded by European Research CouncilAdG Life-his-T to T.N.M.S., a Walter-Hitzig fellowship from theCCI (BMBF 01EO0803), a Deutsche Forschungsgemeinschaft (DFGRO 4120/1-1) research fellowship to J.C.R., a Marie Curie IntraEuropean Fellowship to L.P., a Nederlandse Organisatie voorWetenschappelijk Onderzoek Veni grant (916.86.080) to J.B.B.,and a Leukemia and Lymphoma Society Career DevelopmentFellowship to S.H.N.

Supplementary Materialswww.sciencemag.org/cgi/content/full/science.1235487/DC1Materials and MethodsSupplementary TextFig. S1Table S1References (20–23)

22 January 2013; accepted 7 March 2013Published online 14 March 2013;10.1126/science.1235487

Neural Decoding of VisualImagery During SleepT. Horikawa,1,2 M. Tamaki,1* Y. Miyawaki,3,1† Y. Kamitani1,2‡

Visual imagery during sleep has long been a topic of persistent speculation, but its privatenature has hampered objective analysis. Here we present a neural decoding approach in whichmachine-learning models predict the contents of visual imagery during the sleep-onset period,given measured brain activity, by discovering links between human functional magneticresonance imaging patterns and verbal reports with the assistance of lexical and image databases.Decoding models trained on stimulus-induced brain activity in visual cortical areas showedaccurate classification, detection, and identification of contents. Our findings demonstrate thatspecific visual experience during sleep is represented by brain activity patterns shared bystimulus perception, providing a means to uncover subjective contents of dreaming usingobjective neural measurement.

Dreaming is a subjective experience dur-ing sleep often accompanied by vividvisual contents. Previous research has

attempted to link physiological states with dream-ing (1–3), but none has demonstrated how spe-cific visual contents are represented in brainactivity. The advent of machine-learning–basedanalysis allows for the decoding of stimulus- andtask-induced brain activity patterns to reveal vi-sual contents (4–9).We extended this approach tothe decoding of spontaneous brain activity duringsleep (Fig. 1A). Although dreaming has oftenbeen associated with the rapid-eye movement

(REM) sleep stage, recent studies have demon-strated that dreaming is dissociable from REMsleep and can be experienced during non-REMperiods (10). We focused on visual imagery (hal-lucination) experienced during the sleep-onset(hypnagogic) period (sleep stage 1 or 2) (11, 12)because it allowed us to collect many observa-tions by repeating awakenings and recording par-ticipants’ verbal reports of visual experience.Reports at awakenings in sleep-onset and REMperiods share general features such as frequency,length, and contents, while differing in several as-pects, including the affective component (13–15).We analyzed verbal reports using a lexical data-base to create systematic labels for visual con-tents. We hypothesized that contents of visualimagery during sleep are represented at leastpartly by visual cortical activity patterns sharedby stimulus representation. Thus, we trained de-coders on brain activity induced by natural im-ages from Web image databases.

Three people participated in the functionalmagnetic resonance imaging (fMRI) sleep exper-

iments (Fig. 1A), in which they were woken whenan electroencephalogram signature was detected(16) (fig. S1), and they were asked to give averbal report freely describing their visual ex-perience before awakening [table S1; duration,34 T 19 s (mean T SD)]. We repeated this pro-cedure to attain at least 200 awakenings with avisual report for each participant. On average, weawakened participants every 342.0 s, and visualcontents were reported in over 75% of the awaken-ings (Fig. 1B). Offline sleep stage scoring (fig.S2) further selected awakenings to exclude con-tamination from the wake stage in the periodimmediately before awakening (235, 198, and186 awakenings for participants 1 to 3, respec-tively, used for decoding analyses) (16).

From the collected reports, words describingvisual objects or scenes were manually extractedand mapped to WordNet, a lexical database inwhich semantically similar words are groupedas “synsets” in a hierarchical structure (17, 18)(Fig. 2A). Using a semantic hierarchy, we groupedextracted visual words into base synsets that ap-peared in at least 10 reports from each participant(26, 18, and 16 synsets for participants 1 to 3, re-spectively; tables S2 to S4) (16). The fMRI dataobtained before each awakeningwere labeledwitha visual content vector, each element of whichindicated the presence or absence of a base synsetin the subsequent report (Fig. 2B and fig. S3).Wealso collected images depicting each base synsetfrom ImageNet (19), an image database in whichWeb images are grouped according to WordNet,or from Google Images, for decoder training.

We constructed decoders by training linearsupport vector machines (SVMs) (20) on fMRIdata measured while each person viewed Webimages for each base synset. Multivoxel patternsin the higher visual cortex [HVC; the ventral re-gion covering the lateral occipital complex (LOC),fusiform face area (FFA), and parahippocampalplace area (PPA); 1000 voxels], the lower visual

1ATR Computational Neuroscience Laboratories, Kyoto 619-0288, Japan. 2Nara Institute of Science and Technology, Nara630-0192, Japan. 3National Institute of Information and Com-munications Technology, Kyoto 619-0288, Japan.

*Present address: Brown University, 190 Thayer Street,Providence, RI 02912, USA.†Present address: The University of Electro-Communications,Tokyo 182-8585, Japan.‡Corresponding author. E-mail: [email protected]

www.sciencemag.org SCIENCE VOL 340 3 MAY 2013 639

REPORTS

on

July

10,

201

3w

ww

.sci

ence

mag

.org

Dow

nloa

ded

from

Page 3: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

cortex (LVC; V1 to V3 combined; 1000 voxels),or the subareas (400 voxels for each area) wereused as the input for the decoders (16).

First, a binary classifier was trained on thefMRI responses to stimulus images of two basesynsets (three-volume averaged data correspond-ing to the 9-s stimulus block) and tested on thesleep samples [three-volume (9-s) averaged dataimmediately before awakening] that containedexclusively one of the two synsets while ignoringother concurrent synsets (16) (Fig. 3A). We onlyused synset pairs in which one of the synsets ap-peared in at least 10 reports without co-occurrencewith the other (201, 118, and 86 pairs for par-ticipants 1 to 3, respectively). The distribution ofthe pairwise decoding accuracies for the HVC isshown togetherwith that from the decoders trainedon the same stimulus-induced fMRI data withrandomly shuffled synset labels (Fig. 3B; fig. S4,individual participants). The mean decoding ac-curacy was 60.0%, 95% confidence interval (CI)[(59.0, 61.0%); three participants pooled], which

was significantly higher than that of the label-shuffled decoders with both Wilcoxon rank-sumand permutation tests (P < 0.001).

To look into the commonality of brain activitybetween perception and sleep-onset imagery, wefocused on the synset pairs that produced content-specific patterns in each of the stimulus and sleepexperiments (pairs with high cross-validation clas-sification accuracy within each of the stimulusand sleep data sets; figs. S5 and S6) (16). Withthe selected pairs, even higher accuracies wereobtained [mean = 70.3%, CI (68.5, 72.1); Fig.3B, dark blue; fig. S4, individual participants;tables S5 to S7, lists of the selected pairs], indi-cating that content-specific patterns are highlyconsistent between perception and sleep-onsetimagery. The selection of synset pairs, whichused knowledge of the test (sleep) data, does notbias the null distribution by the label-shuffleddecoders (Fig. 3B, black), because the contentspecificity in the sleep data set alone does notimply commonality between the two data sets.

Additional analyses revealed that the multi-voxel pattern, rather than the average activitylevel, was critical for decoding (figs. S7 and S8).We also found that the variability of decodingperformance among synset pairs can be accountedfor at least partly by the semantic differencesbetween paired synsets. The decoding accuracyfor synsets paired across meta-categories (human,object, scene, and others; tables S2 to S4) wassignificantly higher than that for synsets withinmeta-categories (Wilcoxon rank-sum test, P <0.001; Fig. 3C and fig. S9). However, even with-in a meta-category, the mean decoding accuracysignificantly exceeded chance level, indicatingspecificity to fine object categories.

The mean decoding accuracies for differentvisual areas are shown in Fig. 3D (fig. S10,individual participants). The LVC scored 54.3%,CI (53.4, 55.2) for all pairs, and 57.2%, CI (54.2,60.2) for selected pairs (three participants pooled).The performance was significantly above chancelevel but worse than that for the HVC. Individual

A B

Participant 1

Participant 2

Participant 3

Fig. 1. Experimental overview. (A) fMRI data were acquired from sleeping partic-ipants simultaneously with polysomnography. Participants were awakened during sleepstage 1 or 2 (red dashed line) and verbally reported their visual experience during

sleep. fMRI data immediately before awakening [an average of three volumes (= 9 s)] were used as the input for main decoding analyses (sliding time windowswere used for time course analyses). Words describing visual objects or scenes (red letters) were extracted. The visual contents were predicted using machine-learning decoders trained on fMRI responses to natural images. (B) The numbers of awakenings with and without visual contents are shown for each participant(with numbers of experiments in parentheses).

A BParticipant 2

Fig. 2. Visual content labeling. (A) Words describing visual objects orscenes (red) were mapped onto synsets of the WordNet tree. Synsets weregrouped into base synsets (blue frames) located higher in the tree. (B) Visual reports (participant 2) are represented by visual content vectors, in which thepresence or absence of the base synsets in the report at each awakening is indicated by white or black, respectively. Examples of images used for decoder trainingare shown for some of the base synsets.

3 MAY 2013 VOL 340 SCIENCE www.sciencemag.org640

REPORTS

on

July

10,

201

3w

ww

.sci

ence

mag

.org

Dow

nloa

ded

from

Page 4: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

areas (V1 to V3, LOC, FFA, and PPA) showed agradual increase in accuracy along the visual pro-cessing pathway, mirroring the progressively com-plex response properties from low-level imagefeatures to object-level features (21). When the

time window was shifted, the decoding accuracypeaked around 0 to 10 s before awakening (Fig.3E and fig. S11; no correction for hemodynamicdelay). The high accuracies after awakening maybe due to hemodynamic delay and the large time

window. Thus, verbal reports are likely to reflectbrain activity immediately before awakening.

To read out richer contents given arbitrarysleep data, we next performed a multilabel de-coding analysis in which the presence or absence

A

B C

D

E

Fig. 3. Pairwise decoding. (A) Schematic overview. (B) Distributions ofdecoding accuracies with original and label-shuffled data for all pairs(light blue and gray) and selected pairs (dark blue and black) (threeparticipants pooled). (C) Mean accuracies for the pairs within and acrossmeta-categories (synsets in others were excluded; numbers of pairs are inparentheses). (D) Accuracies across visual areas (numbers of selected pairs

for V1, V2, V3, LOC, FFA, PPA, LVC, and HVC: 45, 50, 55, 70, 48, 78, 55,and 97). (E) Time course (HVC and LVC; averaged across pairs and par-ticipants). The plot shows the performance with the 9-s (three-volume) timewindow centered at each point (gray window and arrow for main analyses).For all results, error bars or shadings indicate 95% CI, and dashed linesdenote chance level.

A

B

C D

E F

Participant 2

Participant 2

Fig. 4. Multilabel decoding. (A) Schematic overview. (B) ROC curves(left) and AUCs (right) are shown for each synset (participant 2; asterisks,Wilcoxon rank-sum test, P < 0.05). (C) AUC averaged within meta-categories for different visual areas (three participants pooled; numbersof synsets in parentheses). (D) Example time course of synset scores for asingle sleep sample (participant 2, 118th; color legend as in (B); reportedsynset, “character,” in bold). (E) Time course of averaged synset scores forreported synsets (red) and unreported synsets with high or low (blue orgray) co-occurrence with reported synsets (averaged across awakeningsand participants). Scores are normalized by the mean magnitude in eachparticipant. (F) Identification analysis. Accuracies are plotted againstcandidate set size for original and extended visual content vectors (av-eraged across awakenings and participants). Because Pearson’s correla-tion coefficient could not be calculated for vectors with identical elements,such samples were excluded. For all results, error bars or shadings in-dicate 95% CI, and dashed lines denote chance level.

www.sciencemag.org SCIENCE VOL 340 3 MAY 2013 641

REPORTS

on

July

10,

201

3w

ww

.sci

ence

mag

.org

Dow

nloa

ded

from

Page 5: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

of each base synset was predicted by a synsetdetector constructed from a combination of pair-wise decoders (Fig. 4A) (16). The synset detectorprovided a continuous score indicating how like-ly the synset is to be present in each report. Wecalculated receiver operating characteristic (ROC)curves for each base synset by shifting the detec-tion threshold for the output score (Fig. 4B, theHVC in participant 2, time window immediatelybefore awakening; fig. S12, all participants), andthe detection performance was quantified by thearea under the curve (AUC). Although the per-formance varied across synsets, 18 out of the total60 synsets were detected with above-chance lev-els (Wilcoxon rank-sum test, uncorrected P <0.05), greatly exceeding the number of synsetsexpected by chance (0.05 × 60 = 3).

Using the AUC, we compared the decodingperformance for individual synsets grouped intometa-categories in different visual areas. Overall,the performance was better in the HVC than inthe LVC, consistent with the pairwise decodingperformance [fig. S13; three participants pooled;analysis of variance (ANOVA), P = 0.003]. Al-though V1 to V3 did not show different perform-ances across meta-categories, the higher visualareas showed a marked dependence on meta-categories (Fig. 4C and fig. S13). In particular,the FFA showed better performance with humansynsets, whereas the PPA showed better perform-ance with scene synsets [ANOVA (interaction),P = 0.001], consistent with the known responsecharacteristics of these areas (22, 23). The LOCand FFA showed similar results, presumably be-cause our functional localizers selected partiallyoverlapping voxels.

The output scores for individual synsets showeddiverse and dynamic profiles in each sleep sam-ple (Fig. 4D, fig. S14, and movies S1 and S2)(16). These profiles may reflect a dynamic varia-tion of visual contents, including those expe-rienced even before the period near awakening.On average, there was a general tendency for thescores for reported synsets to increase toward thetime of awakening (Fig. 4E and fig. S15). Synsetsthat did not appear in reports showed greaterscores if they had a high co-occurrence relation-ship with reported synsets (Fig. 4E; synsets withthe top 15% conditional probabilities given a re-ported synset, calculated from the whole-contentvectors in each participant). The effect of co-occurrence is rather independent of that of se-mantic similarity (Fig. 3C), because both factors(high/low co-occurrence and within/across meta-categories) had highly significant effects on thescores of unreported synsets (timewindow imme-diately before awakening; two-way ANOVA, P <0.001, three participants pooled) with moderateinteraction (P = 0.016). The scores for reportedsynsets were significantly higher than those forunreported synsets even within the same meta-category (Wilcoxon rank-sum test, P < 0.001).Verbal reports are unlikely to describe full detailsof visual experience during sleep, and it is pos-sible that contents with high general co-occurrence

(such as street and car) tend to be experiencedtogether even when all are not reported. There-fore, high scores for the unreported synsets mayindicate unreported but actual visual contentsduring sleep.

Finally, to explore the potential of multilabeldecoding to distinguish numerous contents, weperformed identification analysis (7, 8). Theoutput scores (score vector) were used to identifythe true visual content vector among a variablenumber of candidates (true vector + random vec-tors with matched probabilities for each synset)by selecting the candidate most correlated withthe score vector (repeated 100 times for eachsleep sample to obtain the correct identificationrate) (16). The performance exceeded chance lev-el across all set sizes (Fig. 4F, HVC, three par-ticipants pooled; fig. S16, individual participants),although the accuracies were not as high as thoseachieved using stimulus-induced brain activity inprevious studies (7, 8). The same analysis wasperformed with extended visual content vectorsin which unreported synsets having a high co-occurrence with reported synsets (top 15% con-ditional probability) were assumed to be present.The results showed that extended visual contentvectors were better identified (Fig. 4F and fig.S16), suggesting that multilabel decoding out-puts may represent both reported and unreportedcontents.

Together, our findings provide evidence thatspecific contents of visual experience during sleepare represented by, and can be read out from,visual cortical activity patterns shared with stim-ulus representation. Our approach extends previ-ous research on the (re)activation of the brainduring sleep (24–27) and the relationship be-tween dreaming and brain activity (2, 3, 28) bydiscovering links between complex brain activ-ity patterns and unstructured verbal reports usingdatabase-assistedmachine-learning decoders. Theresults suggest that the principle of perceptualequivalence (29), which postulates a commonneural substrate for perception and imagery, gen-eralizes to spontaneously generated visual expe-rience during sleep. Althoughwe have demonstratedsemantic decoding with the HVC, this does notrule out the possibility of decoding low-levelfeatures with the LVC. The decoding presentedhere is retrospective in nature: Decoders wereconstructed after sleep experiments based on thecollected reports. However, because reported syn-sets largely overlap between the first and the lasthalves of the experiments (59 out of 60 basesynsets appeared in both), the same decodersmay apply to future sleep data. The similaritybetween REM and sleep-onset reports (13–15)and the visual cortical activation during the REMsleep (24, 25, 28) suggest that the same decoderscould also be used to decode REM imagery. Ourmethod may further work beyond the bounds ofsleep stages and reportable experience to uncoverthe dynamics of spontaneous brain activity inassociation with stimulus representation. We ex-pect that it will lead to a better understanding of

the functions of dreaming and spontaneous neu-ral events (10, 30).

References and Notes1. W. Dement, N. Kleitman, J. Exp. Psychol. 53, 339

(1957).2. M. Dresler et al., Curr. Biol. 21, 1833 (2011).3. C. Marzano et al., J. Neurosci. 31, 6674 (2011).4. J. V. Haxby et al., Science 293, 2425 (2001).5. Y. Kamitani, F. Tong, Nat. Neurosci. 8, 679 (2005).6. S. A. Harrison, F. Tong, Nature 458, 632 (2009).7. K. N. Kay, T. Naselaris, R. J. Prenger, J. L. Gallant, Nature

452, 352 (2008).8. Y. Miyawaki et al., Neuron 60, 915 (2008).9. M. Stokes, R. Thompson, R. Cusack, J. Duncan,

J. Neurosci. 29, 1565 (2009).10. Y. Nir, G. Tononi, Trends Cogn. Sci. 14, 88 (2010).11. T. Hori, M. Hayashi, T. Morikawa, in Sleep Onset: Normal

and Abnormal Processes, R. D. Ogilvie, J. R. Harsh, Eds.(American Psychological Association, Washington, DC,1994), pp. 237–253.

12. R. Stickgold, A. Malia, D. Maguire, D. Roddenberry,M. O’Connor, Science 290, 350 (2000).

13. D. Foulkes, G. Vogel, J. Abnorm. Psychol. 70, 231(1965).

14. G. W. Vogel, B. Barrowclough, D. D. Giesler, Arch. Gen.Psychiatry 26, 449 (1972).

15. D. Oudiette et al., Conscious. Cogn. 21, 1129(2012).

16. Materials and methods are available as supplementarymaterials on Science Online.

17. C. Fellbaum, Ed., WordNet: An Electronic LexicalDatabase (MIT Press, Cambridge, MA, 1998).

18. A. G. Huth, S. Nishimoto, A. T. Vu, J. L. Gallant, Neuron76, 1210 (2012).

19. J. Deng et al., IEEE Conference on Computer Vision andPattern Recognition, 248 (2009).

20. V. N. Vapnik, Statistical Learning Theory (Wiley, NewYork, 1998).

21. E. Kobatake, K. Tanaka, J. Neurophysiol. 71, 856(1994).

22. R. Epstein, N. Kanwisher, Nature 392, 598 (1998).23. N. Kanwisher, J. McDermott, M. M. Chun, J. Neurosci. 17,

4302 (1997).24. A. R. Braun et al., Science 279, 91 (1998).25. P. Maquet, J. Sleep Res. 9, 207 (2000).26. Y. Yotsumoto et al., Curr. Biol. 19, 1278 (2009).27. M. A. Wilson, B. L. McNaughton, Science 265, 676

(1994).28. S. Miyauchi, M. Misaki, S. Kan, T. Fukunaga, T. Koike,

Exp. Brain Res. 192, 657 (2009).29. R. A. Finke, Principles of Mental Imagery (MIT Press,

Cambridge, MA, 1989).30. J. A. Hobson, Nat. Rev. Neurosci. 10, 803 (2009).

Acknowledgments: We thank Y. Onuki, T. Beck, Y. Fujiwara,G. Pandey, and T. Kubo for assistance with early experimentsand M. Takemiya and P. Sukhanov for comments on themanuscript. This work was supported by grants from theStrategic Research Program for Brain Science (MEXT), theStrategic Information and Communications R&D PromotionProgramme (SOUMU), the National Institute of Informationand Communications Technology, the Nissan Science Foundation,and the Ministry of Internal Affairs and Communications (Noveland innovative R&D making use of brain structures).

Supplementary Materialswww.sciencemag.org/cgi/content/full/science.1234330/DC1Materials and MethodsFigs. S1 to S16Tables S1 to S7References (31–43)Movies S1 and S2

20 December 2012; accepted 5 March 2013Published online 4 April 2013;10.1126/science.1234330

3 MAY 2013 VOL 340 SCIENCE www.sciencemag.org642

REPORTS

on

July

10,

201

3w

ww

.sci

ence

mag

.org

Dow

nloa

ded

from

Page 6: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

www.sciencemag.org/cgi/content/full/science.1234330/DC1

Supplementary Materials for

Neural Decoding of Visual Imagery During Sleep T. Horikawa, M. Tamaki, Y. Miyawaki, Y. Kamitani*

*Corresponding author. E-mail: [email protected]

Published 4 April 2013 on Science Express

DOI: 10.1126/science.1234330

This PDF file includes:

Materials and Methods Figs. S1 to S16 Tables S1 to S7 References (31–43) Captions for Movies S1 and S2

Other Supplementary Material for this manuscript includes the following: (available at www.sciencemag.org/cgi/content/full/science.1234330/DC1)

Movies S1 and S2

Page 7: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

2

Table of contents Materials and Methods 1. Subjects 2. Prior instructions to subjects 3. Sleep adaptation 4. Sleep experiment 5. MRI acquisition 6. PSG recordings 7. Offline EEG artifact removal and sleep-stage scoring 8. Visual content labeling 9. Visual stimulus experiment 10. Localizer experiments 11. MRI data preprocessing 12. Region of interest (ROI) selection 13. Decoding analysis 14. Synset pair selection by within-dataset cross-validation Supplementary Text for Online Movies Supplementary Figures 1. Time course of theta power 2. Time course of sleep state proportion 3. Visual content vectors 4. Distribution of pairwise decoding accuracies 5. Stimulus-to-stimulus pairwise decoding 6. Sleep-to-sleep pairwise decoding 7. Time course of BOLD signals 8. Decoding with averaged vs. multivoxel activity 9. Decoding within and across meta-categories 10. Pairwise decoding accuracies across visual cortical areas 11. Time course of pairwise decoding accuracy 12. ROC analysis for all subjects 13. AUCs for meta-categories in different visual areas 14. Examples for the time courses of synset scores 15. Time courses of averaged synset scores 16. Identification analysis Supplementary Tables 1. Examples of verbal reports 2–4. List of base synsets for Subject 1–3 5–7. Pairwise decoding performance of the selected pairs for Subject 1–3

Page 8: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

3

Materials and Methods The study protocol was approved by the Ethics Committee of ATR. 1. Subjects Potential subjects answered questionnaires regarding their sleep-wake habits. Usual sleep and wake times, regularity of the sleep habits, habits of taking a nap, sleep complaints, and regularity of lifestyle (e.g., mealtime), their physical and psychiatric health, and sleeping conditions were checked. Anyone who had physical or psychiatric diseases, was currently receiving medical treatment or was suspected of having a sleep disorder was excluded. People who had a habit of taking alcoholic beverages before sleep or smoking were also excluded. Finally, three healthy subjects (Japanese-speaking male adults, aged 27–39 years) with normal visual acuity participated in the experiments. All subjects gave written informed consent for their participation in the experiment. 2. Prior instructions to subjects From three days prior to each experiment, subjects were instructed to maintain their sleep-wake habits, i.e., daily wake/sleep time and sleep duration. They were also instructed to refrain from excessive alcohol consumption, unusual physical exercise, and taking of naps, from the day before each experiment. Their sleep-wake habits were monitored by a sleep log. No subject was chronically sleep deprived, and all slept over 6 hours on average the night before experiments. 3. Sleep adaptation Subjects underwent two adaptation sleep experiments before the main fMRI sleep experiments to get used to sleeping in the experimental setting (31, 32). The adaptation experiments were conducted using the same procedures as the fMRI sleep experiments except that real fMRI scans were not performed. The experimental environment was simulated using a mock scanner consisting of the shell of a real scanner without the magnet. Echo-planar imaging (EPI) acoustic noise was also simulated and given to the subject via speakers. 4. Sleep experiment Sleep (nap) experiments were carried out from 1:00 pm until 5:30 pm, and were scheduled to include the mid-afternoon dip (33). Subjects were instructed to sleep if they could, but not to try to sleep if they felt they could not. This instruction was given to reduce psychological pressure toward sleeping because efforts to sleep may themselves cause difficulty in falling asleep. fMRI scans were conducted simultaneous with PSG recordings (electroencephalogram [EEG], electrooculogram [EOG], electromyogram [EMG], and electrocardiogram [ECG]). We performed multiple awakenings (see below for details) to collect verbal reports on visual experience during sleep. The multiple-awakening procedure was repeated while fMRI was performed continuously (interrupted upon the subject’s request for a break; mean ± SD of duration across all 55 runs, 88.99 ± 26.09 min [mean ± SD]). The experiment was repeated over 10, 7, and 7 days in Subject 1–3, respectively, until at least 200 awakenings with a visual report were obtained from each subject. Offline sleep stage scoring confirmed that in >90% of the awakenings followed by visual reports, the last 15-s epoch before awakening was classified as sleep stage 1 or 2 (fig. S2). If the last 15-s epoch was classified as the waking stage, we excluded the data from decoding analysis. As a result, 235, 198, and 186 awakenings were selected for Subject 1–3, respectively, constituting sleep data samples for further analysis. Multiple-awakening procedure. Once the fMRI scan began, the subject was allowed to fall asleep. The experimenter monitored noise-reduced EEG recordings in real time while performing EEG staging on every 6-s epoch. The experimenter awakened subjects by calling them by name via a microphone/headphone when a single epoch with alpha-wave suppression and theta-wave (ripple) occurrence, which are known to be EEG signatures of NREM sleep stage 1 for obtaining frequent visual reports upon awakening (11), was detected (see fig. S1 for theta profiles). The subject was asked to verbally describe what they saw before awakening along with other mental content and then to go to sleep again. If the EEG signatures were detected before the elapsed time from the previous awakening reached 2 min, then the experimenter waited until it reached 2–3 min. If the subject was already in NREM sleep stage 2 when 2 min had passed, and it was unlikely they would go back to NREM sleep stage 1, the subject was awakened. When the subject repeatedly entered NREM sleep stage 2 within 2 min, the subject was

Page 9: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

4

awakened with short intervals (less than 2 min) or was asked to remain awake with eyes closed for one or two minutes after the awakening to increase their arousal level. This multiple-awakening procedure was repeated during the fMRI session. The subject was also asked to respond by button press when they detected a beep sound (250 Hz pure tone, 500 ms duration, 12–18-s inter-stimulus intervals). This auditory task was conducted for potential use in monitoring the sleep stage. However, in the present study, we did not use the data because we failed to record responses in some of the experiments owing to computer trouble. Our preliminary analysis using successfully recorded data (Subject 1) showed that the detection rates in each of the wake/sleep stages were similar to those of previous work (34) even when sleep samples were limited to those in which a sound was played during the last 15-s epoch before awakening but not detected by the subject, the decoding results were similar. Subjects were informed that they could quit the experiment anytime they wished to, and that they could refuse to report mental contents in cases where there were privacy concerns. Acquisition of verbal reports. On each awakening, the subject was asked if the subject had seen anything just before awakening, and then to freely describe it along with other mental contents. If a description of the contents was unclear, the subject was asked to report the contents in more detail. Most reports started immediately upon awakening and lasted for 34 ± 19 s (mean ± SD; three subjects pooled). After the free verbal report, the subject was asked to answer specific questions such as rating the vividness of the image and the subjective timing of the experience (from when and until when relative to awakening), but the reports obtained by these explicit questions were not used in the analyses in the current study. Free reports that contained at least one visual element were classified as visual reports. If no visual content was present, reports were classified as others including thought (active thinking), forgot, non-visual report, and no report. The classification was first conducted in real time by the experimenter, and was later confirmed by other investigators. Examples of verbal reports are shown in table S1. The subject’s voice during the procedure was recorded by an optical microphone. 5. MRI acquisition fMRI data were collected using 3.0-Tesla scanner located at the ATR Brain Activity Imaging Center. An interleaved T2*-weighted gradient-EPI scan was performed to acquire functional images to cover whole brain (sleep experiments, visual stimulus experiments, and higher visual area localizer experiments: TR, 3,000 ms; TE, 30 ms; flip angle, 80 deg; FOV, 192 × 192 mm; voxel size, 3 × 3 × 3 mm; slice gap, 0 mm; number of slices, 50) or the entire occipital lobe (retinotopy experiments; TR, 2,000 ms; TE, 30 ms; flip angle, 80 deg; FOV, 192 × 192 mm; voxel size, 3 × 3 × 3 mm; slice gap, 0 mm; number of slices, 30). T2-weighted turbo spin echo images were scanned to acquire high-resolution anatomical images of the same slices used for the EPI (sleep experiments, visual stimulus experiments, and higher visual area localizer experiments: TR, 7,020 ms; TE, 69 ms; flip angle, 160 deg; FOV, 192 × 192 mm; voxel size, 0.75 × 0.75 × 3.0 mm; retinotopy experiments: TR, 6,000 ms; TE, 57 ms; flip angle, 160 deg; FOV, 192 × 192 mm; voxel size, 0.75 × 0.75 × 3.0 mm). T1-weighted magnetization-prepared rapid acquisition gradient-echo (MP-RAGE) fine-structural images of the whole head were also acquired (TR, 2,250 ms; TE, 3.06 ms; TI, 900 ms; flip angle, 9 deg, FOV, 256 × 256 mm; voxel size, 1.0 × 1.0 × 1.0 mm). 6. PSG recordings PSG was performed simultaneously with fMRI. PSG consisted of EEG, EOG, EMG, and ECG recordings. EEGs were recorded at the 31 scalp sites in all experiments except for one (Fp1, Fp2, F7, F3, Fz, F4, F8, FC5, FC1, FC2, FC6, T7, C3, Cz, C4, T8, TP9, CP5, CP1, CP2, CP6, TP10, P7, P3, Pz, P4, P8, POz, O1, Oz, O2) according to 10% electrode positions (35). For one experiment, EEG was recorded at 25 scalp sites (Fp1, Fp2, F7, F3, Fz, F4, F8, T7, C3, Cz, C4, T8, P7, P3, Pz, P4, P8, PO7, PO3, POz, PO4, PO8, O1, Oz, O2). EOGs were recorded bipolarly from four electrodes placed at the outer canthi of both eyes (horizontal EOG) and above and below the right eye (vertical EOG). EMGs were recorded bipolarly from the mentum. ECGs were recorded from the lower shoulder blade. EEG and ECG recordings were referenced to FCz. Electrode impedance was kept below 20 kΩ. All data were recorded by an MRI-compatible amplifier at a sampling rate of 5,000 Hz using BrainVision Recorder. The artifacts derived from T2*-weighted gradient-EPI scan and ballistocardiogram (36) were reduced in real time using RecView so that the experimenter was able to monitor the EEG patterns online. Because EEG recordings were referenced to FCz, EEGs recorded in the occipital area were better than those recorded in the central area at detecting subtle changes in EEG waves. Thus, O1 was used for online monitoring of EEG data. When the O1 channel was contaminated by artifacts, O2 was used instead.

Page 10: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

5

7. Offline EEG artifact removal and sleep-stage scoring Artifacts were removed offline from EEG recordings after each experiment (the FMRIB plug-in for EEGLAB, The University of Oxford) for further analyses. EEG data were then down-sampled to 500 Hz, re-referenced to TP9 or TP10, depending on the derivation, and low-pass filtered with a 216-Hz cut-off frequency using a two-way least-squares FIR filter. EEG recordings were scored into sleep stages according to the standard criteria (37) for every 15 s. 8. Visual content labeling The subject’s report at each awakening, given verbally in Japanese, was transcribed into text. The reports that contained at least one visual object or scene were classified as visual report, and those without visual content were classified as others. Three labelers extracted words (nouns) that described visual objects or scenes from each visual report text and translated them into English (verified by a bilingual speaker). These words were mapped to WordNet, a lexical database in which words with a similar meaning are grouped as synsets in a hierarchical structure (17). Synset assignment was cross-checked by all three labelers. For each extracted word, hypernym synsets of the assigned synset were also identified in the WordNet hierarchy. To determine representative synsets that describe visual contents, we selected the synsets (synsets assigned to each word and their hypernyms) that were found in 10 or more visual reports without co-occurrence with at least one other synset. We then removed the synsets that were the hypernyms of others. These procedures produced base synsets, which were frequently reported while being semantically exclusive and specific. The visual contents at each awakening were coded by a visual content vector, in which the presence/absence of each synset was denoted by 1/0. 9. Visual stimulus experiment We selected stimulus images for decoder construction using ImageNet (http://www.image-net.org/; 2011 fall release) (19), an image database in which web images are grouped according to WordNet. Two-hundred and forty images were collected for each base synset (each image corresponding to exclusively one synset). If images for a synset were not available in ImageNet, we collected images using Google Images (http://www.google.com/imghp). In this experiment, the selected images were resized so that the width and height of images were within 16 deg (the original aspect ratio was preserved) and were presented at the center of the screen on gray background. Subjects were allowed to freely view the images without fixation. We measured stimulus-induced fMRI activity for each base synset by presenting these images (using the same fMRI setting as the sleep experiment). In a 9-s stimulus block, six images were randomly sampled from the 192 (Subject 1) or 240-image (Subject 2 and Subject 3) set for one base synset without replacement, and each image was presented for 0.75 s with intervening blanks of 0.75 s. Thus each of the images for each base synset was presented only once during the experiment. The stimulus block (9 s) was followed by a 6-s rest period, and was repeated for all base synsets in each run. Extra 33-s and 6-s rest periods were added at the beginning and the end of each run, respectively. We repeated the run with different images to obtain 32, 40, and 40 blocks per base synset for Subject 1–3, respectively. 10. Localizer experiments Retinotopy. The retinotopy mapping session followed the conventional procedure (38, 39) using a rotating wedge and an expanding ring of a flickering checkerboard. The data were used to delineate the borders between each visual cortical area, and to identify the retinotopic map on the flattened cortical surfaces of individual subjects. Localizers for higher visual areas. We performed functional localizer experiments to identify the lateral occipital complex (LOC), fusiform face area (FFA), and parahippocampal place area (PPA) for each individual subject (22, 23, 40). The localizer experiment consisted of 4–8 runs and each run contained 16 stimulus blocks. In this experiment, intact or scrambled images (12 × 12 deg) of face, object, house and scene categories were presented at the center of the screen. Each of eight stimulus types (four categories × two conditions) was presented twice per run. Each stimulus block consisted of a 15-s intact or scrambled stimulus presentation. The intact and scrambled stimulus blocks were presented successively (the order of the intact and scrambled stimulus block was random), followed by a 15-s rest period of uniform gray background. Extra 33-s and 6-s rest periods were added at the beginning and end of each run, respectively. In each stimulus block, twenty different images of the same type were presented with a duration of 0.3 s followed by intervening blanks of 0.4-s duration. Images for each category were collected from the following resources: face images from THE CENTER FOR VITAL LONGEVITY (http://vitallongevity.utdallas.edu) (41); object images from The Object Databank (http://stims.cnbc.cmu.edu/ImageDatabases/TarrLab/Objects/; Stimulus images courtesy of Michael J. Tarr, Center

Page 11: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

6

for the Neural Basis of Cognition and Department of Psychology Carnegie Mellon University, http://www.tarrlab.org/); house and scene images from SUN database (http://groups.csail.mit.edu/vision/SUN/) (42). 11. MRI data preprocessing The first 9-s scans for experiments with TR = 3 s (sleep, visual stimulus, and higher visual area localizer experiments) and 8-s scans for experiments with TR = 2 s (retinotopy experiments) of each run were discarded to avoid instability of the MRI scanner. The acquired fMRI data underwent three-dimensional motion correction by SPM5 (http://www.fil.ion.ucl.ac.uk/spm). The data were then coregistered to the within-session high-resolution anatomical image of the same slices used for EPI and subsequently to the whole-head high-resolution anatomical image. The coregistered data were then reinterpolated by 3 × 3 × 3 mm voxels. For the sleep data, after a linear trend was removed within each run, voxel amplitude around awakening was normalized relative to the mean amplitude during the period 60–90 s prior to each awakening. The average proportions of wake, stage 1, and stage 2 during this period were 32.7%, 44.0%, and 23.3%, respectively (three subjects pooled). This period was used as the baseline because it tended to show relatively stable BOLD signals over time. We assumed that the occurrence of sleep-onset imagery would be rare with the relatively low theta amplitudes (fig. S1) (11). However, it cannot be ruled out that visual imagery may have been experienced during this period, and that the baseline using this period as the baseline would make it difficult to detect visual contents relevant to the imagery in this period. The voxel values averaged across the three volumes (9 s) immediately before awakening served as a data sample for decoding analysis (the time window was shifted for time course analysis). For the visual stimulus data, after within-run linear trend removal, voxel amplitudes were normalized relative to the mean amplitude of the pre-rest period of each run and then averaged within each 9-s stimulus block (three volumes) shifted by 3 s (one volume) to compensate for hemodynamic delays. Voxels used for the decoding of a synset pair in each ROI were selected by t statistics comparing the mean responses to the images of paired synsets (highest absolute t values; 400 voxels for individual areas, and 1,000 voxels for LVC and HVC). The voxel values in each data sample were z-transformed for removing potential mean-level differences between the sleep and visual stimulus experiments. 12. Region of interest (ROI) selection V1, V2, and V3 were delineated in the standard retinotopy experiment (38, 39), and the lateral occipital complex (LOC), the fusiform face area (FFA), and the parahippocampal place area (PPA) were identified using conventional functional localizers (22, 23, 40). The retinotopy experiment data were transformed to the Talairach coordinates and the visual cortical borders were delineated on the flattened cortical surfaces using BrainVoyager QX (http://www.brainvoyager.com). The voxel coordinates around the gray-white matter boundary in V1–V3 were identified and transformed back into the original coordinates of the EPI images. The voxels from V1, V2, and V3 were combined as the lower visual cortex (LVC; ~2,000 voxels in each subject). The localizer experiment data for higher visual areas were analyzed using SPM5. The voxels that showed significantly higher activation in response to objects, faces, or scenes than scrambled images for each (t test, uncorrected P < 0.05 or 0.01) were identified, and were used as ROIs for LOC, FFA, and PPA, respectively. We set relatively low thresholds for identifying ROIs so that a larger number of voxels would be included to preserve broadly distributed patterns. A continuous region covering LOC, FFA, and PPA was manually delineated on the flattened cortical surfaces, and the region was defined as the higher visual cortex (HVC; ~2,000 voxels in each subject). Voxels overlapping with LVC were excluded from HVC. After voxels of extremely low signal amplitudes were removed, approximately 2,000 voxels remained in LVC (2054, 2172, and 1935 voxels for Subject 1–3, respectively) and in HVC (1956, 1788, and 2235 voxels for Subject 1–3, respectively). For the analysis of individual subareas, the following numbers of voxels were identified for V1, V2, V3, LOC, FFA, and PPA, respectively: 885, 901, 728, 523, 537, and 353 voxels for Subject 1; 779, 949, 897, 329, 382, and 334 voxels for Subject 2; 710, 859, 765, 800, 432, and 316 voxels for Subject 3. LOC, FFA, and PPA identified by the localizer experiments may overlap: the voxels in the overlapping region were included in both ROIs. 13. Decoding analysis For all pairs of the base synsets, a binary decoder consisting of linear Support Vector Machine (20) (SVM; implemented by LIBSVM (43)) was trained on the visual stimulus data of each ROI. The fMRI signals of the selected voxels and the synset labels were given to the decoder as training data. SVM provided a linear discriminant function for classification between synset k and l given input voxel values x = [ x1 , …, xD ] ( D , number of voxels),

Page 12: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

7

fkl (x) = wd xd + w0d=1

D

where wd is the weight parameter for voxel d, and w0 is the bias. The performance was evaluated by the correct classification rate for all sleep samples selected for each pair. The synsets of a pair can have unbalanced numbers of samples, which could potentially lead to some bias. However, when the correct rates were calculated in each synset and then averaged, the averaged correct rates were highly correlated with the correct rates for all pooled samples (correlation coefficients for Subject 1–3, 0.96, 0.97, and 0.97, respectively). Therefore the bias, if any, is likely to be small. In the pairwise decoding analysis, the stimulus-trained decoder was tested on the sleep data that contained exclusively one of the paired synsets. Prediction was made on the basis of whether fkl (x) was positive ( k ) or negative ( l ), given a sleep fMRI data sample. In the multilabel decoding, the discriminant functions comparing a base synset k and each of the other synsets ( l ≠ k ) were averaged after the normalization by the norm of the weight vector wkl to yield the linear detector function (5), which indicates how likely synset k is to be present,

fk (x) =1

N −1fkl (x)

|| wkl ||l≠k∑

where N is the number of base synsets. Given a sleep fMRI data sample, multilabel decoding produced a vector consisting of the output scores of the detector functions for all base synsets [ f1(x) , f2 (x) , …, fN (x) ]. 14. Synset pair selection by within-dataset cross-validation To select synset pairs with content-specific patterns in both the stimulus-induced and sleep datasets, we performed cross-validation decoding analysis for each pair in each dataset. For the stimulus-induced dataset, samples from one run were left for testing, and the rest were used for decoder training (repeated until all runs were tested; leave-one-run-out cross validation) (5). For the sleep dataset, one sample was left for testing, and the rest were used for decoder training (leave-one-sample-out cross-validation). Note that since the frequency in the sleep reports is generally different between paired synsets, their available samples are usually unbalanced. To avoid possible biases in decoder training caused by the imbalance, we trained multiple decoders by randomly resampling a subset of the training data for the synset with more samples to match to the synset with fewer samples (repeated 11 times). The discriminant functions calculated for all the resampled training datasets were averaged after normalization by the norm of the weight vector to yield the discriminant function (decoder) to be used for testing in each cross-validation step. We selected the synset pairs that showed high cross-validation performance in both of datasets (one-tailed binomial test, uncorrected P < 0.05).

Page 13: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

8

Supplementary Text for Online Movies The movies illustrate the evolution of contents predicted by multilabel decoding for two individual sleep samples. The image panel displays stimulus images averaged with the weights according to the scores of the synsets at each time point. Stimulus images were randomly picked from the image samples for each synset. The graph shows the time courses of individual synset scores (red, reported synsets; black, unreported synsets). The tag clouds at the sides illustrate the names of the base synsets whose sizes vary according to the synset scores. The verbal reports for these sleep samples are as follows (note that the reports were originally Japanese):

Movie S1. “What I was just looking at was some kind of characters. There was something like a writing paper for composing an essay, and I was looking at the characters from the essay or whatever it was. It was in black and white and the writing paper was the only thing that was there. And shortly before that I think I saw a movie with a person in it or something but I can't really remember.” Movie S2. “Well, there were persons, about 3 persons, inside some sort of hall. There was a male, a female, and maybe like a child. Ah, it was like a boy, a girl, and a mother. I don't think that there was any color.”

The report for Movie S1 contains three words describing visual objects or scene (character, writing paper, and person), and only character was assigned with a base synset (character synset [ID:06818970], character). The report for Movie S2 contains eight visual words (person, hall, male, female, child, boy, girl, and mother), and five out of the eight words were assigned with base synsets (male synset [ID:09624168], male, boy; female synset [ID:09619168], female, girl, mother). Note that several visual words were not assigned with base synsets, and thus, were not reflected in the movies (see footnote of table S1).

Page 14: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

9

Subject 1Subject 2Subject 3

Time to awakening (s)

0

1N

orm

aliz

ed th

eta

pow

er (a

.u)

120 90 60 30 0

Baseline

Fig. S1. Time course of theta power. The time course of theta power (4–7 Hz) during 2 min before awakening is shown for each subject (error bar, 95% CI; averaged across awakenings). For each awakening, we shifted the 9-s window (equivalent to three fMRI volumes) by 3 s (equivalent to one fMRI volume) to calculate the theta power from preprocessed EEG signals. The plotted points indicate the center of the 9-s window (slightly displaced to avoid overlaps between subjects). The power was normalized relative to the mean power during the time window of 60–90 s prior to each awakening (gray area). The result of this offline analysis is consistent with the awakening procedure in which theta ripples were detected in online monitoring to determine the awakening timing.

Page 15: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

10

0

50

100

No visual contentWith visual contentSubject 1

0

50

100 Subject 2

0

50

100 Subject 3

Prop

ortio

n (%

)

Time to awakening (s)120 60 0 120 60 0

Wake Stage1 Stage2 Fig. S2. Time course of sleep state proportion. The proportion of wake/sleep states (Wake, Stage 1, and Stage 2) is shown for all the awakenings with and without visual content in each subject. Offline sleep stage scoring was conducted for every 15-s epoch using simultaneously recorded EEG data. The last 15-s epoch before awakening was classified as Sleep Stage 1 or 2 in over 90% of the awakenings with visual contents, but in fewer awakenings with no visual content. This result suggests that most visual reports are indeed associated with imagery during sleep, not with imagery during wakefulness. The samples in which the last 15-s epoch before awakening was classified as Wake were not used for further analyses.

Page 16: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

11

Subject 1

Subject 2

Subject 3

Awakening index

50 100 150 200

50 100 150

50 100 150

buildingchair

characterclothing

codecognition

external body partgeographical area

girlgroup

illustrationimplement

linemale

materialnatural object

performerpicture

roomshapetable

vertebrateway

windowworkplace

writing

bookbuilding

carcharacter

commoditycomputer screen

coveringdwelling

electronic equipmentfemale

foodfurniture

malemercantile establishment

pointregion

representationstreet

carcommunication

displayexternal body part

femalegeological formation

housemale

materialroom

surfacetabletract

vascular plantvertebrate

way

Base

syn

sets

Fig. S3. Visual content vectors. Visual content vectors are shown for all subjects and awakenings with visual reports (excluding those with contamination of the wake stage detected by offline sleep staging). Each column denotes the presence/absence of the base synsets in the sleep sample. Note that there are several samples in which no synset is present. This is because the reported words in these samples were rare and not included in the base synsets.

Page 17: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

12

20 8050

All (86 pairs)

Selected (11 pairs)

20 8050

All (118 pairs)

Selected (47 pairs)

Shuffled

Unshuffled

Decoding

accuracy (%)

20 8050

All (201 pairs)

Selected (39 pairs)

Chance

Subject 3HVC

Subject 2HVC

Subject 1HVC

Fig. S4. Distribution of pairwise decoding accuracies. The distribution of the stimulus-to-sleep pairwise decoding accuracies for the higher visual cortex (HVC) is shown together with that from label-shuffled decoders. The format is the same as in Fig. 3B. The mean decoding accuracies for all pairs were 56.0% (95% CI [54.7, 57.3]) for Subject 1, 66.9% (CI [65.1, 66.8]) for Subject 2, and 59.8% (CI [57.9, 61.7]) for Subject 3, and those for selected pairs were 65.3% (CI [62.4, 68.2]) for Subject 1, 75.4% (CI [73.3, 77.4]) for Subject 2, and 66.1% (CI [61.2, 71.0]) for Subject 3. The fraction of pairs that showed significantly better decoding performance than chance level (one-tailed binomial test, uncorrected P < 0.05) was 22.4% (45/201) for Subject 1, 57.6% (68/118) for Subject 2, and 27.9% (24/86) for Subject 3. The performance was significantly better than that from label-shuffled decoders for all subjects for both of all and selected pairs (Wilcoxon rank-sum test, P < 0.001).

Page 18: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

13

20 8050

Subject 3HVC

20 8050

Subject 2HVC

Shuffled

Unshuffled

Decoding

accuracy (%)

20 8050

Subject 1HVC

Ch

an

ce

Fig. S5. Stimulus-to-stimulus pairwise decoding. The cross-validation analysis of the stimulus-induced dataset (see Materials and Methods “14. Synset pair selection by within-dataset cross-validation”) yielded the distribution of accuracies for stimulus-to-stimulus pairwise decoding (Subject 1–3, HVC; shown together with the distribution from label-shuffled decoders). The mean decoding accuracies were 85.6% (95% CI [84.1, 87.0]) for Subject 1, 87.0% (CI [85.4, 88.7]) for Subject 2, and 81.4% (CI [79.4, 83.4]) for Subject 3. The fraction of pairs that showed significantly better decoding performance than chance level (one-tailed binomial test, uncorrected P < 0.05) was 94.5% (190/201) for Subject 1, 99.8% (117/118) for Subject 2, and 95.3% (82/86) for Subject 3. The performance was significantly better than that from label-shuffled decoders for all subjects (Wilcoxon rank-sum test, P < 0.001 for all three subjects).

Page 19: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

14

20 8050

Subject 3HVC

20 8050

Subject 2HVC

Shuffled

Unshuffled

Decoding

accuracy (%)

20 8050

Subject 1HVC C

ha

nce

Fig. S6. Sleep-to-sleep pairwise decoding. The cross-validation analysis of the sleep dataset (see Materials and Methods “14. Synset pair selection by within-dataset cross-validation”) yielded the distribution of accuracies for sleep-to-sleep pairwise decoding (Subject 1–3, HVC; shown with the distribution from label-shuffled decoders). The mean decoding accuracies were 52.9% (95% CI [51.2, 54.7]) for Subject 1, 60.1% (CI [58.1, 62.1]) for Subject 2, and 51.3% (CI [49.1, 53.5]) for Subject 3. The fraction of pairs that showed significantly better decoding performance than chance level (one-tailed binomial test, uncorrected P < 0.05) was 20.4% (41/201) for Subject 1, 40.7% (48/118) for Subject 2, and 12.8% (11/86) for Subject 3. The results showed that the performance was significantly better than that from label-shuffled decoders for two of three subjects (Wilcoxon rank-sum test, P = 0.002 for Subject 1; P < 0.001 for Subject 2; P = 0.302 for Subject 3). Note that although the sleep-to-sleep decoding shows marginally significant accuracies (depending on subjects), it is not as accurate as the stimulus-to-sleep decoding. This is presumably because training samples from the sleep dataset were fewer and noisier than those from the stimulus-induced dataset, and thus decoders were not well trained with the sleep dataset.

Page 20: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

15

ï

% s

igna

l cha

nge

Subject 2

ï

Subject 3

Baseline

ï

Time to awakening (s)

Subject 1With visual contentNo visual content

Fig. S7. Time course of BOLD signals. The time course of the averaged BOLD signal in HVC is shown for the awakenings with and without visual content (shades, 95% CI; mean across awakenings). The gray area indicates the period used for the baseline for the normalization. The mean activity level in HVC did not show marked differences toward the timing of awakening whether the subjects reported visual contents or not.

Page 21: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

16

Multivo

xel

Averag

ed

Multivo

xel

Averag

ed

Multivo

xel

Averag

ed

50

80

Dec

odin

g ac

cura

cy (%

)

Subject 1 Subject 2 Subject 3

Fig. S8. Decoding with averaged vs. multivoxel activity. The voxel values in each data sample from HVC were averaged (before the z-transformation in each sample), and the averaged activity was used as the input to the pairwise decoders. The decoding accuracy with averaged activity is compared with that from the multivoxel decoders (error bar, 95% CI; averaged across all pairs). The performance with averaged activity is close to the chance level and significantly worse than that with multivoxel activity (Wilcoxon signed-rank test, P < 0.001, all subjects).

Page 22: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

17

50

80

Within

(12)

Across

(54)

Dec

odin

g ac

cura

cy (%

)

Subject 1

Within

(26)

Across

(124

)

Subject 2

Within

(10)

Across

(40)

Subject 3

Fig. S9. Decoding within and across meta-categories. Pairwise decoding accuracies grouped into the pairs within and across meta-categories are shown for individual subjects. The numbers of available pairs are denoted in parentheses. Because the others category contains a large variety of synsets in terms of semantic similarity, the pairs of synsets in the others category were excluded from this analysis. The performance of across-meta-category decoding was better than that of within-meta-category decoding in all subjects, though the significance level varied (Wilcoxon rank-sum test, P = 0.261 for Subject 1; P = 0.003 for Subject 2; P = 0.020 for Subject 3; error bar, 95% CI).

Page 23: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

18

50

80

50

80 AllSelected

LVC HVC V1 V2 V3 LOC FFA PPAArea

50

80De

codi

ng a

ccur

acy

(%)

Subject 3

Subject 1

Subject 2

Fig. S10. Pairwise decoding accuracies across visual cortical areas. The decoding accuracies with different visual areas are shown for individual subjects (error bars, 95% CI). Selected pairs were determined on the basis of the cross-validation analysis in each area (the numbers of selected pairs for V1, V2, V3, LOC, FFA, PPA, LVC, and HVC: 24, 22, 29, 38, 36, 22, 27, and 39 pairs for Subject 1; 15, 23, 21, 24, 7, 47, 20, and 47 pairs for Subject 2; 6, 5, 5, 8, 5, 9, 8, and 11 pairs for Subject 3, respectively).

Page 24: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

19

Time to awakening (s)

ï ï36 0 ï ï36 0

HVC LVC

Subject 3

50

Subject 1

50

D

eco

din

g a

ccu

racy (

%) Subject 2

50

All

Selected

Fig. S11. Time course of pairwise decoding accuracy. The time course of pairwise decoding accuracy is shown for individual subjects (shades, 95% CI; averaged across all or selected pairs). Averages of three fMRI volumes (9 s; HVC or LVC) around each time point were used as inputs to the decoders. The performance is plotted at the center of the window. The gray region indicates the time window used for the main analyses (the arrow denotes the performance obtained from the time window). No corrections for hemodynamic delay were conducted. Note that fMRI signals after awakening may be contaminated with movement artifacts and brain activity associated with mental states during verbal reports. Mental states during verbal reports are unlikely to explain the high accuracy immediately after awakening, because the accuracy profile does not match the mean duration of verbal reports (34 ± 19 s, mean ± SD; three subjects pooled).

Page 25: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

20

table:0.520clothing:0.528window:0.537chair:0.548implement:0.552

male:0.477performer:0.707girl:0.724

way:0.490

building:0.628workplace:0.620

room:0.693JHRJUDSKLFDOïDUHD

QDWXUDOïREMHFWH[WHUQDOïERG\ïSDUWcode:0.444material:0.455illustration:0.470group:0.486FRJQLWLRQZULWLQJOLQH

YHUWHEUDWHpicture:0.658

character:0.644

shape:0.805

0 0.2 0.4 0.6 0.8 0

0.2

0.4

0.6

0.8

False positive

Subject 3

0

0.2

0.4

0.6

0.8

Subject 1

0

0.2

0.4

0.6

0.8

True

pos

itive

Subject 2

Human

Object Others

Scene

female:0.647PDOH

representation:0.448character:0.767

IRRGFRPSXWHUïVFUHHQfurniture:0.562commodity:0.596FDUFRYHULQJ

electronic-equipment:0.700

book:0.776dwelling:0.605point:0.647building:0.664

mercantile-establishment:0.760

street:0.774region:0.794

Human

Object

Others

Scene

table:0.475display:0.506car:0.624

way:0.539house:0.545WUDFWroom:0.600JHRORJLFDOïIRUPDWLRQIHPDOH

male:0.688

YDVFXODUïSODQWmaterial:0.486H[WHUQDOïERG\ïSDUWvertebrate:0.632communication:0.673VXUIDFH

Human

ObjectOthers

Scene

Fig. S12. ROC analysis for all subjects. Above-chance level detection (Wilcoxon rank-sum test, uncorrected P < 0.05; denoted by *) was achieved in 7/26 synsets for Subject 1, 8/18 for Subject 2, and 3/16 for Subject 3. Similar results were obtained using the samples with no visual report to represent the absence of each synset (4/26 synsets for Subject 1, 7/18 for Subject 2, and 4/16 for Subject 3).

Page 26: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

21

LVCHVC

V3

V1V2

LOCFFAPPA

Object

(5)

Human

(3)

Scene

(5)

Others

(13)

Object

(8)

Human

(2)

Scene

(6)

Others

(2)

Object

(16)

Human

(7)

Scene

(16)

Others

(21)

Object

(3)

Human

(2)

Scene

(5)

Others

(6)

0.5

0.7

0.5

0.7

0.5

0.7

AUC

Subject 1 Subject 2 Subject 3 Pooled

Fig. S13. AUCs for meta-categories in different visual areas. The mean AUC for the synsets within each meta-category is plotted for different visual areas (individual subjects’ and pooled results; error bars, 95% CI). The numbers of available synsets are shown in parentheses. Because the sample sizes are small in individual subjects, evaluation with statistical tests is difficult. However, tendencies similar to the pooled results are found in individual subjects: HVC tends to outperform LVC (ANOVA, P = 0.029 for Subject 1, P = 0.159 for Subject 2, and P = 0.139 for Subject 3), and FFA and PPA tend to show better performances with human and scene, respectively (ANOVA [interaction], P = 0.099 for Subject 1; P = 0.028 for Subject 2; P = 0.044 for Subject 3). Additionally, FFA and LOC showed similar results in individual subjects.

Page 27: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

22

Subject 1202th awakening

Scor

e

Time to awakening (s)

Subject 2144th awakening

ï

0

10

ï

0

10

female

male

male

48 36 24 12 0 ï ï

Subject 334th awakening

ï

0

10 display

Fig. S14. Examples for the time courses of synset scores. The time courses of synset scores from multilabel decoding analyses are shown for four individual sleep examples. The plots represent the scores for the reported synset(s) (bold line with synset name) and the unreported synsets using the colors in the legend for fig. S12.

Page 28: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

23

48 36 24 12 0 ï ï

Score

Time to awakening (s)

Subject 2

Subject 3

Subject 1ReportedUnreported (High/Low)

ï

0

ï

0

ï

0

Fig. S15. Time courses of averaged synset scores. Synset scores were averaged across awakenings for reported (red) and unreported synsets with high (blue) and low (gray) co-occurrence in individual subjects (shades, 95% CI).

Page 29: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

24

0

20

40

60

80 Subject 1

0

20

40

60

80 Subject 2

2 4 8 16 32

0

20

40

60

80

Candidate set size

Subject 3

Ide

ntifica

tio

n a

ccu

racy (

%)

Original

Extended

True

N candidates

Decoder

output

Most similar vector?

BA

Fig. S16. Identification analysis. (A) The correlation coefficients between the score vector from multilabel decoding and each of the candidate vectors consisting of the true visual content vector and a variable number of binary vectors generated randomly with a matched probability were calculated. The vector with the highest correlation coefficient was chosen as the one representing the visual contents. (B) Identification performance with original and extended visual content vectors is shown for individual subjects.

Page 30: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

25

Table S1. Examples of Verbal Reports

Subject

1

Subject

2

Subject

3

IndexSubject Report Word Base synset

13 Well, what was that? Two male persons, well, what was that? I cannot

remember very well, but there were e-mail texts. There were also

characters. E-mail address? Yes, there were a lot of e-mail

addresses. And two male persons existed.

male person

text

character

male (male person)

character (grapheme,

graphic symbol)

writing (written material,

piece of writing)

130 Yes, ah, yes, a female person, well, existed. The person served some

foods, umm, like a flight attendant. Then, well, before that scene, I

saw a scene in which I ate or saw yogurt, or I saw yogurt or a scence

in which yogurt was served. What appeared was the female person

and an unknown thing like a refrigerator. Maybe indoor, with colors.

food

flight attendant

yogurt

refrigerator

person

female person

female (female person)

commodity (trade good,

good)

food (solid food)

114 Well, from the sky, from the sky, well, what was it? I saw something

like a bronze statue, a big bronze statue. The bronze statue existed

on a small hill. Below the hill, there were houses, streets, and trees in

an ordinary way.

sky

statue

hill

house

street

tree

geological formation

(formation)

house

way

vascular plant

(tracheophyte)

133 Well, somewhere, in a place like a studio to make a TV program or

something, well, a male person ran with short steps, or run, from the

left side to the right side. Then, he tumbled. He stumbled over

something, and stood up while laughing, and said something. He said

something to persons on the left side. So, well, a person ran, and

there were a lot of unknown people. I did not saw female persons.

There were a group of a lot of people, either male or female. The

place was like a studio. Though there are a lot of variety for studios,

the studio was a huge room. Since it was a huge room, it was indoor

maybe. I saw such a scene in the huge room.

studio

room

male

person

people

54 Yes, I had a dream. Something long. First at some shop. Ah, a bakery

shop. I was choosing some merchandise. I took a roll in which a leaf

of perilla was put. Then, I went out, and on the street, I saw a person

who were doing something like taking a photograph.

shop

bakery

merchandise

roll

leaf

perilla

street

person street

commodity (trade good,

good)

food (solid food)

point

mercantile establishment

(retail store, sales outlet,

outlet)

186 Well, in the night, somewhere, well, in a restaurant in office building

covered with windowpanes, on the big table, someone important

looked a menu and chose dishes. There were both male and female

persons. Then, there was a night scenery from the window.

restaurant

office building

windowpane

table

someone

menu

male

female

scenery

window

table

communication

male (male person)

female (female person)

male (male person)

room

workplace (work)

group (grouping)

Note. Originally, the reports and words were verbally reported in Japanese. They were transcribed into text and were translated into English (verified by a

bilingual speaker). Note that not all reported visual words were assigned with a base synset (e.g., person in report #133 of Subject 1, and perilla in report #54

of Subject 2). This is because 1) the word and its hypernyms did not appear in ten or more reports, or 2) the synset assigned to the word is a hypernym of

other synsets (see Materials and Methods “8. Visual content labeling”). On average, 47.7% of reported visual words were assigned with base synsets (49.5%,

50.0%, and 42.0% for Subject 1–3, respectively).

Page 31: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

26

Base synset ID Definition CountReported word Meta-categoryHumanmale (male person) 09624168 a person who belongs to the sex that

cannot have babies127gentleman, boy, middle-aged man,

old man, young man, male, dandy

character (grapheme, graphic symbol)

Others06818970 a written symbol that is used to represent speech

35character, letter

room Scene04105893 an area within a building enclosed by walls and floor and ceiling

20booth, conference room, room, toilet

workplace (work) Scene04602044 a place where work is done 17laboratory, recording studio, studio, workplace

external body part Others05225090 any body part visible externally 17lip, hand, face

natural object Others00019128 an object occurring naturally; not made by man

13leaf, branch, figure, beard, mustache, orange, coconut, moon, sun

building (edifice) Scene02913152 a structure that has a roof and walls and stands more or less permanently in one place

12bathhouse, building, house, restaurant, schoolhouse, school

clothing (article of clothing, vesture, wear, wearable, habiliment)

Object03051540 a covering designed to be worn on a person's body

13clothes, baseball cap, clothing, coat, costume, tuxedo, silk hat, hat, T-shirt, kimono, muffler, polo shirt, suit, uniform

chair Object03001627 a seat for one person, with a support for the back

12chair, folding chair, wheelchair

picture (image, icon, ikon) Others03931044 a visual representation (of an object or scene or person or abstraction) produced on a surface

12graphic, picture, image, portrait

shape (form) Others00027807 the spatial arrangement of something as distinct from its substance

11circle, square, quadrangle, box, node, point, dot, tree, tree diagram, hole

vertebrate (craniate) Others01471682 animals having a bony or cartilaginous skeleton with a segmented spinal column and a large brain enclosed in a skull or cranium

11bird, ostrich, raptor, hawk, falcon, eagle, frog, snake, dog, leopard, horse, sheep, monkey, fish, skipjack tuna

implement Object03563967 instrumentation (a piece of equipment or tool) used to effect an end

10trigger, hammer, ice pick, pan, pen, pencil, plunger, pole, pot, return key, stick, wok

way Scene04564698 any artifact consisting of a road or path affording passage from one place to another

11hallway, hall, passageway, pedestrian crossing, stairway, street

window Object04588739 (computer science) a rectangular part of a computer screen that contains a display different from the rest of the screen

11window

girl (miss, missy, young lady, young woman, fille)

Human10129825 a young woman 11girl, young woman

material (stuff) Others14580897 the tangible substance that goes into the makeup of a physical object

11water, paper, sand, wood, sheet, leaf, page

cognition (knowledge, noesis) Others00023271 the psychological result of perception and learning and reasoning

10symbol, profile, monster, character

group (grouping) Others00031264 any number of entities (members) considered as a unit

10string, people, pair, pop group, band, calendar, line, forest

table Object04379243 a piece of furniture having a smooth flat top that is usually supported by one or more vertical legs

10desk, stand, table

code (computer code) Others06355894 (computer science) the symbolic arrangement of data or instructions in a computer program or the set of such instructions

10computer code, code

writing (written material, piece of writing)

Others06362953 the work of a writer; anything expressed in letters of the alphabet (especially when considered from the point of view of style and effect)

10draft, text, document, written document, clipping, line

line Others06799897 a mark that is long relative to its width 10line

illustration Others06999233 artwork that helps make something clear or attractive

10illustration, figure, Figure

geographical area (geographic area, geographical region, geographic region)

Scene08574314 a demarcated area of the Earth 10tennis court, campus, playing field, ground, field, lawn, park, parking area, square, public square, town

performer (performing artist) Human10415638 an entertainer who performs a dramatic or musical work for an audience

10idol, singer, actor, actress, clown, comedian

Note. In this study, a representative instance provided by WordNet for each synset was used as the name of the base synset. Here, other instances were described

in parentheses.

Table S2. List of Base Synsets for Subject 1

Page 32: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

27

character (grapheme, graphic symbol)

06818970 a written symbol that is used to represent speech

34character Others

male (male person) 09624168 a person who belongs to the sex that cannot have babies

27boy, male, male person Human

street 04334599 a thoroughfare (usually including sidewalks) that is lined with buildings

21street Scene

car (auto, automobile, machine, motorcar)

02958343 a motor vehicle with four wheels; usually propelled by an internal combustion engine

17car, patrol car, police cruiser, used-car

Object

food (solid food) 07555863 any solid substance (as opposed to liquid) that is used as a source of nourishment

19food, chocolate bar, apple pie, cake, cookie, bread, roll, noodle, tomato, cherry tomato, yogurt, yoghurt

Object

building (edifice) 02913152 a structure that has a roof and walls and stands more or less permanently in one place

18apartment house, apartment building, building, coffee shop, house, library, school

Scene

representation Others04076846 a creation that is a visual or tangible rendering of someone or something

14map, model, photograph, photo, picture, snowman

furniture (piece of furniture, article of furniture)

03405725 furnishings that make a room or other area ready for occupancy

13 Objectbed, chair, counter, desk, furniture, hospital bed, sofa, couch, table

female (female person) 09619168 a person who belongs to the sex that can have babies

13 Humangirl, wife, female, female person

book (volume) 02870092 physical objects consisting of a number of pages bound together

12 Objectbook, notebook

point 08620061 the precise location of something; a spatially limited location

12 Scenebakery, corner, crossing, intersection, crossroad, laboratory, level crossing, studio, bus stop, port, Kobe

commodity (trade good, good)

03076708 articles of commerce 11 Objecthat, iron, jacket, T-shirt, Kimono, kimono, merchandise, refrigerator, shirt, stove

computer screen (computer display)

03085602 a screen used to display the output of a computer to the user

10 Objectcomputer screen, computer display

electronic equipment 03278248 equipment that involves the controlled conduction of electrons (especially in a gas or vacuum or semiconductor)

11 Objectamplifier,mobile phone, cell phone, cellular phone, printer, television, TV

mercantile establishment(retail store, sales outlet, outlet)

03748162 a place of business for retailing goods 11 Scenebakery, bookstore, booth, convenience store, department store, shopping center, shopping mall, shop, stall, supermarket

region 08630985 a large indefinite location on the surface of the Earth

10 Scenegarden, downtown, park, parking area, scenery, town, Kobe

covering 03122748 an artifact that covers something else (usually to protect or shelter or conceal it)

10 Objectaccessory, accessories, clothes, covering, flying carpet, hat, jacket, T-shirt, Kimono, kimono, shirt, slipper

dwelling (home, domicile, abode, habitation, dwelling house)

03259505 housing that someone is living in 10 Scenehome,house

Base synset ID Definition CountReported word Meta-category

Table S3. List of Base Synsets for Subject 2

Page 33: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

28

28male (male person) 09624168  a person who belongs to the sex that cannot have babies

old man, male Human

Sceneway 04564698 any artifact consisting of a road or path affording passage from one place to another

19entrance, hallway, footpath, penny arcade, pipe, sewer, staircase, stairway, street, tunnel

Sceneroom 04105893 an area within a building enclosed by walls and floor and ceiling

lobby, hospital room, kitchen, trunk, operating room, room

18

Scene08673395 an extended area of land 18garden, field, ground, athletic field, green, lawn, grassland, rice paddy, paddy field, park, parking area, savannah, savanna

tract (piece of land, piece of ground, parcel of land, parcel)

Human14a person who belongs to the sex that can have babies

09619168female (female person) female

Otherscommunication 00033020 something that is communicated by orto or between people or groups

13subtitle, text, menu, theatre ticket, poster, character, traffic light, traffic signal, graph, mark

Othersvertebrate (craniate) 01471682 animals having a bony or cartilaginous skeleton with a segmented spinal column and a large brain enclosed in a skull or cranium

12bird, hawk, eagle, frog, whale, dog, cheetah, horse, water buffalo, sheep, giraffe, fish

Objectdisplay (video display) 03211117 an electronic device that represents information in visual form

11computer screen, computer display, display, screen, monitor, window

Otherssurface 04362025 the outer boundary of an artifact or a material layer constituting or resembling such a boundary

11ceiling, floor, platform, screen, stage

Othersmaterial (stuff) 14580897 the tangible substance that goes into the makeup of a physical object

12gravel, cardboard, clay, earth, water, log, paper, playing card

Objectcar (auto, automobile, machine, motorcar)

02958343 a motor vehicle with four wheels; usually propelled by an internal combustion engine

11car, jeep, sport car, sports car, sportcar

Scenehouse 03544360 a dwelling that serves as living quarters for one or more families

12house

Othersexternal body part 05225090 any body part visible externally 11head, neck, chest, breast, leg, foot, hand, finger, face, human face, face

Scenegeological formation (formation)

09287968 (geology) the geological features of the earth

11bank, beach, gorge, hill, mountain, slope

Objecttable 04379243 a piece of furniture having a smooth flat top that is usually supported by one or more vertical legs

10counter, desk, operating table, table

Scenevascular plant (tracheophyte)

13083586 green plant having a vascular system: ferns, gymnosperms, angiosperms

10flower, rice, tree

Base synset ID Definition CountReported word Meta-category

Table S4. List of Base Synsets for Subject 3

Page 34: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

29

Within/AcrossP value% correct % correct % correct

0.6442246.9 68.895.3 Acrossexternal body part vs room

0.6062547.1 70.662.5 Withincognition vs vertebrate

0.5953047.4 68.467.2 Acrossperformer vs vertebrate

0.5838148.0 80.079.7 Acrosswindow vs workplace

0.3273655.0 80.0100.0 Acrossperformer vs code

0.2952152.3 57.7100.0 Acrossshape vs male

0.2131957.1 78.684.4 Acrosswindow vs room

0.2078057.7 69.295.3 Acrossvertebrate vs workplace

0.1327460.0 76.787.5 Acrossmaterial vs room

0.1196661.5 69.296.9 Acrossgirl vs workplace

0.0813363.0 66.781.3 Acrosspicture vs workplace

0.0813363.0 85.295.3 Acrossvertebrate vs room

0.0722065.2 73.992.2 Acrossgirl vs chair

0.0720663.3 73.396.9 Acrossperformer vs room

0.0549566.7 81.082.8 Acrossvertebrate vs chair

0.0530064.5 80.795.3 Acrossgirl vs room

0.0445865.5 72.490.6 Acrosswriting vs room

0.0428668.8 75.095.3 Withincognition vs shape

0.0283968.0 72.095.3 Acrossshape vs workplace

0.0208472.2 76.092.2 Withinshape vs picture

0.0157665.9 65.996.9 Acrossgirl vs character

0.0131573.7 68.479.7 Acrossshape vs clothing

0.0128370.4 74.195.3 Acrossillustration vs room

0.0126775.0 80.096.9 Acrossperformer vs table

0.0118370.0 83.385.9 Acrosspicture vs room

0.0106866.0 66.098.4 Acrossworkplace vs character

0.0085171.4 78.687.5 Acrossshape vs room

0.0054577.8 72.2100.0 Acrossperformer vs window

0.0033179.0 89.5100.0 Acrossgirl vs shape

0.0011383.3 88.9100.0 Acrossperformer vs shape

0.0000082.0 68.095.3 Acrossroom vs character

0.0005385.7 76.298.4 Acrosstable vs girl

0.0419968.4 68.498.4 Acrossgirl vs window

0.0497168.4 68.498.4 Acrossperformer vs writing

0.9060238.7 67.798.4 Acrossexternal body part vs workplace

vertebrate vs natural object 0.0033179.0 68.473.4 Within

0.0404463.2 63.271.9 Withinmaterial vs character

0.0327870.6 76.598.4 Acrossperformer vs illustration

0.0467762.5 62.584.4 Acrosschair vs character

P <

0.05

22/3

9 pa

irsP

> 0.

0517

/39

pairs

Synset pairStimulus-to-sleep Stimulus-to-stimulus Sleep-to-sleep

Table S5. Pairwise Decoding Performance of the Selected Pairs for Subject 1

Note. Performance of stimulus-to-sleep, stimulus-to-stimulus, and sleep-to-sleep decoding using HVC is shown for the pairs whose stimulus-to-stimulus and

sleep-to-sleep decoding accuracies had P < 0.05 (uncorrected one-tailed binomial test). The fMRI activity from the 9 s before each awakening was used as

input for the stimulus-to-sleep and sleep-to-sleep decoding analyses. The order of the pairs is sorted in the ascending order of P values for stimulus-to-sleep

decoding.

Page 35: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

30

Within/AcrossSynset pair

P <

0.05

42/4

7 pa

irsP

> 0.

055/

47 p

airs

P value% correctStimulus-to-sleep

% correctStimulus-to-stimulus

Withinelectronic equipment vs food 0.1552259.3 81.3 66.7Withindwelling vs mercantile establishment 0.1542661.1 66.3 72.2Acrossregion vs representation 0.0679865.2 66.3 78.3Acrossregion vs commodity 0.0633266.7 95.0 81.0Acrosscomputer screen vs male 0.0588962.9 92.5 68.6Acrossfemale vs street 0.0452464.5 98.8 67.7

Acrosselectronic equipment vs building 0.0416366.7 77.5 66.7Acrossfood vs male 0.0413762.8 96.3 67.4Acrossrepresentation vs car 0.0363266.7 90.0 70.4Withindwelling vs region 0.0327870.6 67.5 70.6Acrossmercantile establishment vs female 0.0206170.8 100.0 70.8

Acrossregion vs computer screen 0.0170773.7 78.8 68.4Acrossrepresentation vs building 0.0147969.2 66.3 65.4

Acrossfood vs street 0.0090569.7 96.3 69.7Acrossdwelling vs book 0.0065176.5 93.8 70.6Withincommodity vs car 0.0054274.1 90.0 77.8Acrossmercantile establishment vs male 0.0050671.4 100.0 80.0Acrossstreet vs male 0.0039970.0 98.8 77.5Acrosscomputer screen vs street 0.0036774.1 86.3 66.7Withinfood vs car 0.0031972.7 97.5 69.7Acrossregion vs electronic equipment 0.0029880.0 86.3 75.0

Acrosselectronic equipment vs street 0.0023475.0 87.5 67.9Withinelectronic equipment vs car 0.0021976.9 80.0 73.1Acrossmercantile establishment vs electronic equipment 0.0018481.0 77.5 76.2Withinrepresentation vs character 0.0018373.0 83.8 70.3Acrosscommodity vs character 0.0015272.1 92.5 62.8

Withincovering vs car 0.0009180.0 91.3 72.0Acrossbook vs male 0.0009075.0 95.0 69.4Acrossregion vs book 0.0005185.0 93.8 80.0Acrosscar vs male 0.0002676.2 82.5 78.6Acrossfurniture vs character 0.0002375.6 98.8 64.4

Acrossbook vs building 0.0001882.6 88.8 69.6Acrosscommodity vs street 0.0000685.2 93.8 66.7Acrosspoint vs character 0.0000479.1 98.8 69.8Acrossmercantile establishment vs book 0.0000294.7 88.8 79.0Acrossbook vs street 0.0000285.7 92.5 75.0Withinbook vs car 0.0000288.5 92.5 76.9Acrossdwelling vs character 0.0000182.1 96.3 79.5

Acrossfemale vs character 0.0000082.2 95.0 71.1Acrossmale vs character 0.0000080.4 97.5 67.9Acrossmercantile establishment vs character 0.0000086.1 95.0 79.1Acrossstreet vs character 0.0000083.3 97.5 79.2Acrossbuilding vs character 0.0000084.4 96.3 80.0Acrosscar vs character 0.0000091.3 92.5 78.3

% correct

Acrosscovering vs street 0.0002381.5 98.8 66.7

Acrosscovering vs character 0.0010673.2 98.8 68.3

0.0452464.5 Acrossfurniture vs street 71.3 64.5

Sleep-to-sleep

Table S6. Pairwise Decoding Performance of the Selected Pairs for Subject 2

Page 36: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

31

0.7192645.2 64.565.0 Withintract vs room

0.0014272.5 62.592.5 Acrosstract vs male

0.0038771.1 71.186.3 Acrossmaterial vs male

0.0090569.7 81.895.0 Acrossroom vs male

0.0159067.7 67.792.5 Acrosscommunication vs male

0.0317366.7 66.782.5 Acrosscommunication vs room

0.3393453.9 65.467.5 Withingeological formation vs room

0.0178871.4 71.483.8 Acrossgeological formation vs communication

0.0316864.9 67.692.5 Acrossdisplay vs male

0.0082475.0 70.071.3 Withinmaterial vs communication

0.0209269.2 69.290.0 Acrossfemale vs room

Within/AcrossSynset pair P value% correctStimulus-to-sleep

% correctStimulus-to-stimulus

% correctSleep-to-sleep

P <

0.05

9/11

pai

rsP

> 0.

052/

11 p

airs

Table S7. Pairwise Decoding Performance of the Selected Pairs for Subject 3

Page 37: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

1

References and Notes 1. W. Dement, N. Kleitman, The relation of eye movements during sleep to dream activity: An

objective method for the study of dreaming. J. Exp. Psychol. 53, 339 (1957). doi:10.1037/h0048189 Medline

2. M. Dresler et al., Dreamed movement elicits activation in the sensorimotor cortex. Curr. Biol. 21, 1833 (2011). doi:10.1016/j.cub.2011.09.029 Medline

3. C. Marzano et al., Recalling and forgetting dreams: Theta and alpha oscillations during sleep predict subsequent dream recall. J. Neurosci. 31, 6674 (2011). doi:10.1523/JNEUROSCI.0412-11.2011 Medline

4. J. V. Haxby et al., Distributed and overlapping representations of faces and objects in ventral temporal cortex. Science 293, 2425 (2001). doi:10.1126/science.1063736 Medline

5. Y. Kamitani, F. Tong, Decoding the visual and subjective contents of the human brain. Nat. Neurosci. 8, 679 (2005). doi:10.1038/nn1444 Medline

6. S. A. Harrison, F. Tong, Decoding reveals the contents of visual working memory in early visual areas. Nature 458, 632 (2009). doi:10.1038/nature07832 Medline

7. K. N. Kay, T. Naselaris, R. J. Prenger, J. L. Gallant, Identifying natural images from human brain activity. Nature 452, 352 (2008). doi:10.1038/nature06713 Medline

8. Y. Miyawaki et al., Visual image reconstruction from human brain activity using a combination of multiscale local image decoders. Neuron 60, 915 (2008). doi:10.1016/j.neuron.2008.11.004 Medline

9. M. Stokes, R. Thompson, R. Cusack, J. Duncan, Top-down activation of shape-specific population codes in visual cortex during mental imagery. J. Neurosci. 29, 1565 (2009). doi:10.1523/JNEUROSCI.4657-08.2009 Medline

10. Y. Nir, G. Tononi, Dreaming and the brain: From phenomenology to neurophysiology. Trends Cogn. Sci. 14, 88 (2010). doi:10.1016/j.tics.2009.12.001 Medline

11. T. Hori, M. Hayashi, T. Morikawa, in Sleep Onset: Normal and Abnormal Processes, R. D. Ogilvie, J. R. Harsh, Eds. (American Psychological Association, Washington, 1994), pp. 237–253.

12. R. Stickgold, A. Malia, D. Maguire, D. Roddenberry, M. O’Connor, Replaying the game: Hypnagogic images in normals and amnesics. Science 290, 350 (2000). doi:10.1126/science.290.5490.350 Medline

13. D. Foulkes, G. Vogel, Mental activity at sleep onset. J. Abnorm. Psychol. 70, 231 (1965). doi:10.1037/h0022217 Medline

14. G. W. Vogel, B. Barrowclough, D. D. Giesler, Limited discriminability of REM and sleep onset reports and its psychiatric implications. Arch. Gen. Psychiatry 26, 449 (1972). doi:10.1001/archpsyc.1972.01750230059012 Medline

15. D. Oudiette et al., Dreaming without REM sleep. Conscious. Cogn. 21, 1129 (2012). doi:10.1016/j.concog.2012.04.010 Medline

16. Materials and methods are available as supplementary materials on Science Online.

Page 38: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

2

17. C. Fellbaum, Ed., WordNet: An Electronic Lexical Database (MIT Press, Cambridge, MA, 1998)

18. A. G. Huth, S. Nishimoto, A. T. Vu, J. L. Gallant, A continuous semantic space describes the representation of thousands of object and action categories across the human brain. Neuron 76, 1210 (2012). doi:10.1016/j.neuron.2012.10.014 Medline

19. J. Deng et al., Imagenet: A large-scale hierarchical image database. IEEE CVPR (2009)

20. V. N. Vapnik, Statistical Learning Theory (Wiley, New York, 1998)

21. E. Kobatake, K. Tanaka, Neuronal selectivities to complex object features in the ventral visual pathway of the macaque cerebral cortex. J. Neurophysiol. 71, 856 (1994). Medline

22. R. Epstein, N. Kanwisher, A cortical representation of the local visual environment. Nature 392, 598 (1998). doi:10.1038/33402 Medline

23. N. Kanwisher, J. McDermott, M. M. Chun, The fusiform face area: A module in human extrastriate cortex specialized for face perception. J. Neurosci. 17, 4302 (1997). Medline

24. A. R. Braun et al., Dissociated pattern of activity in visual cortices and their projections during human rapid eye movement sleep. Science 279, 91 (1998). doi:10.1126/science.279.5347.91 Medline

25. P. Maquet, Functional neuroimaging of normal human sleep by positron emission tomography. J. Sleep Res. 9, 207 (2000). doi:10.1046/j.1365-2869.2000.00214.x Medline

26. Y. Yotsumoto et al., Location-specific cortical activation changes during sleep after training for perceptual learning. Curr. Biol. 19, 1278 (2009). doi:10.1016/j.cub.2009.06.011 Medline

27. M. A. Wilson, B. L. McNaughton, Reactivation of hippocampal ensemble memories during sleep. Science 265, 676 (1994). doi:10.1126/science.8036517 Medline

28. S. Miyauchi, M. Misaki, S. Kan, T. Fukunaga, T. Koike, Human brain activity time-locked to rapid eye movements during REM sleep. Exp. Brain Res. 192, 657 (2009). doi:10.1007/s00221-008-1579-2 Medline

29. R. A. Finke, Principles of Mental Imagery (MIT Press, Cambridge, MA, 1989)

30. J. A. Hobson, REM sleep and dreaming: towards a theory of protoconsciousness. Nat. Rev. Neurosci. 10, 803 (2009). Medline

31. H. W. Agnew Jr., W. B. Webb, R. L. Williams, The first night effect: an EEG study of sleep. Psychophysiology 2, 263 (1966). doi:10.1111/j.1469-8986.1966.tb02650.x Medline

32. M. Tamaki, H. Nittono, M. Hayashi, T. Hori, Examination of the first-night effect during the sleep-onset period. Sleep 28, 195 (2005). Medline

33. T. H. Monk, D. J. Buysse, C. F. Reynolds, 3rd, D. J. Kupfer, Circadian determinants of the postlunch dip in performance. Chronobiol. Int. 13, 123 (1996). doi:10.3109/07420529609037076 Medline

34. R. D. Ogilvie, R. T. Wilkinson, The detection of sleep onset: Behavioral and physiological convergence. Psychophysiology 21, 510 (1984). doi:10.1111/j.1469-8986.1984.tb00234.x Medline

Page 39: Neural Decoding of Visual Imagery During Sleep T. Horikawa et al. … · 2013-07-10 · DOI: 10.1126/science.1234330 Science 340, 639 (2013); T. Horikawa et al. Neural Decoding of

3

35. F. Sharbrough et al., American Electroencephalographic Society guidelines for standard electrode position nomenclature. J. Clin. Neurophysiol. 8, 200 (1991). doi:10.1097/00004691-199104000-00007 Medline

36. G. Bonmassar et al., Motion and ballistocardiogram artifact removal for interleaved recording of EEG and EPs during MRI. Neuroimage 16, 1127 (2002). doi:10.1006/nimg.2002.1125 Medline

37. A. Rechtschaffen, A. Kales, A Manual of Standardized Terminology, Techniques and Scoring System for Sleep Stages of Human Subjects (U.S. Dept. of Health, Education, and Welfare, Public Health Services-National Institutes of Health, National Institute of Neurological Diseases and Blindness, Neurological Information Network, Bethesda, MD, 1968)

38. S. A. Engel et al., fMRI of human visual cortex. Nature 369, 525 (1994). doi:10.1038/369525a0 Medline

39. M. I. Sereno et al., Borders of multiple visual areas in humans revealed by functional magnetic resonance imaging. Science 268, 889 (1995). doi:10.1126/science.7754376 Medline

40. Z. Kourtzi, N. Kanwisher, Cortical regions involved in perceiving object shape. J. Neurosci. 20, 3310 (2000). Medline

41. M. Minear, D. C. Park, A lifespan database of adult facial stimuli. Behav. Res. Methods Instrum. Comput. 36, 630 (2004). doi:10.3758/BF03206543 Medline

42. J. Xiao, J. Hays, K. Ehinger, A. Oliva, A. Torralba, SUN Database Large scale Scene Recognition from Abbey to Zoo. IEEE CVPR (2010)

43. C. C. Chang, C. J. Lin, LIBSVM a library for support vector machines. ACM Transactions on Intelligent Systems and Technology 2 (2011). Software available at http://www.csie.ntu.edu.tw/~cjlin/libsvm.