Advertisement
Research Article

Scale-Free Brain-Wave Music from Simultaneously EEG and fMRI Recordings

  • Jing Lu,

    Affiliation: Key Laboratory for NeuroInformation of Ministry of Education, School of Life Science and Technology, University of Electronic Science and Technology of China, Chengdu, China

    X
  • Dan Wu,

    Affiliation: Key Laboratory for NeuroInformation of Ministry of Education, School of Life Science and Technology, University of Electronic Science and Technology of China, Chengdu, China

    X
  • Hua Yang,

    Affiliations: Key Laboratory for NeuroInformation of Ministry of Education, School of Life Science and Technology, University of Electronic Science and Technology of China, Chengdu, China, Sichuan Conservatory of Music, Chengdu, China

    X
  • Cheng Luo,

    Affiliation: Key Laboratory for NeuroInformation of Ministry of Education, School of Life Science and Technology, University of Electronic Science and Technology of China, Chengdu, China

    X
  • Chaoyi Li,

    Affiliations: Key Laboratory for NeuroInformation of Ministry of Education, School of Life Science and Technology, University of Electronic Science and Technology of China, Chengdu, China, Center for Life Sciences, Shanghai Institutes for Biological Sciences, Chinese Academy of Sciences, Shanghai, China

    X
  • Dezhong Yao mail

    dyao@uestc.edu.cn

    Affiliation: Key Laboratory for NeuroInformation of Ministry of Education, School of Life Science and Technology, University of Electronic Science and Technology of China, Chengdu, China

    X
  • Published: November 14, 2012
  • DOI: 10.1371/journal.pone.0049773

Abstract

In the past years, a few methods have been developed to translate human EEG to music. In 2009, PloS One 4 e5915, we developed a method to generate scale-free brainwave music where the amplitude of EEG was translated to music pitch according to the power law followed by both of them, the period of an EEG waveform is translated directly to the duration of a note, and the logarithm of the average power change of EEG is translated to music intensity according to the Fechner's law. In this work, we proposed to adopt simultaneously-recorded fMRI signal to control the intensity of the EEG music, thus an EEG-fMRI music is generated by combining two different and simultaneous brain signals. And most importantly, this approach further realized power law for music intensity as fMRI signal follows it. Thus the EEG-fMRI music makes a step ahead in reflecting the physiological process of the scale-free brain.

Introduction

Music and language define us as human [1]. Emotional expression and communication, through language or non-linguistic artistic expression, are recognized as being strongly linked to health and sense of well-being [2]. Therefore, as an artistic expression, music may represent human mind or mood.

In 1934, Adrian and Matthews attempted to listen to the brainwave signals via an amplified speaker [3]. During the past decades, scientists and artists found many methods to make an electroencephalogram(EEG) sonification, although it is difficult for composition to balance music principles and EEG features [4]. Meanwhile, in order to learn more about ourselves, researchers also used the deoxyribonucleic acid(DNA) [5], proteins [6], electromyograms (EMGs) [7] to compose music in the last century.

From the 1990s, various new EEG music generating rules were created [8]. One of them was to translate some parameters of EEG to the parameters of music [9], and another one was to utilize some characteristics such as the epileptic discharges to trigger specific music [10], or to link brain states to various music pieces through Brain Computer Interface [4].

In 2009, we proposed a method to translate EEG to music. The translation rules included the direct mapping from the period of an EEG waveform to the duration of a note, the logarithmic mapping of the average power (AP) change of EEG to music intensity according to the Fechner's law [11], and a scale-free based mapping from the amplitude of EEG to music pitch according to the power law [12]. However, in this method, the pitch and intensity were not independent enough under the translation rules as both pitch and intensity are related to EEG amplitude, so that the music was not strictly in accordance with the composition regulation in which pitch and intensity are usually not mutually related. Meanwhile, the intensity of music usually follows power law [13], however, the intensity of our previous brainwave music was obtained from the AP change of EEG within a time window, it didn't obey the power law (Shown in the following Figure 9). In this work, in order to imitate the general music composition better, we selected another brain information to represent the intensity instead of the EEG amplitude. As the intrinsic metabolic functional activities based functional magnetic resonance imaging(fMRI) is widely used to study the operational organization of the human brain [14], and fortunately, the fMRI blood oxygenation level dependent(BOLD) signal does follow the power law [15], thus the currently widely adopted fMRI may provide us a potential information for intensity of brainwave music. In fact, the fMRI BOLD signal is indirectly related to the electrical activities of a group of neurons by neuro-vascular coupling relation, thus it may reflect the brain mental state.

Data and Methods

1 Ethics Statement

This study was approved by the Research Ethics Board at University of Electronic Science and Technology of China. All participants were asked to read and sign an informed consent form before participating in the study. After experiment, all participants received monetary compensation for their time and effort.

2 EEG-fMRI Data

For the simultaneous EEG-fMRI recordings, the subjects were a 31-year-old female (Subject A) and a 14-year-old female (Subject B). They were both in resting state and scanned in a 3T MRI scanner (EXCITE, GE Milwaukee, USA).

For composing music, the EEG recordings were re-referenced to zero with a software called REST developed in our laboratory [16], [17]. In this work, we chose the EEG at Cz electrode for brainwave music composition, which is at the central of the head and is a channel less affected by the body movement, and took the fMRI signal at the MNI(Montreal Neurological Institute) [18] coordinate (15,−48, 60), which was just below the electrode position Cz. In this way, we assumed the EEG and fMRI signals were almost from the same neural mass. The signals used in this work are illustrated in Figure 1.

thumbnail

Figure 1. The EEG and fMRI signals.

(a) and (d) the fMRI signals collected at MNI coordinate (15,−48, 60); (b) and (e) pieces of signals in (a) and (d) for the following music compositions; (c) and (f) 30 s simultaneous EEGs collected at scalp Cz, respectively for Subjects A and B.

doi:10.1371/journal.pone.0049773.g001

3 EEG-fMRI brain Music

Music note consists of four characters, pitch, timbre, duration and intensity. In this work, we paid special attention to pitch and intensity. And timbre was fixed with piano, which could be changed according to person's hobbies,while the duration was determined by the period of an EEG waveform.

3.1 Pitch

In this study, we still adopt the power law rule between the amplitude (Amp) of an EEG waveform and the Pitch of a musical note [12],(1)
In equation (1), b is the maximum value of all pitches. Parameter a denotes the scale characteristics, and it is determined by the following detrended fluctuation analysis (DFA) [19], [20].

For a discrete time series , the first step of DFA is to subtract the mean from the series and then create a new series by integration:(2)
where denotes the mean of the time series. Next, seriesis divided into a number of segments with length k (k represents the time scale of observation). For each of these segments, a local least-squares linear fit is conducted, and the resulted piece wise linear fit function is designated. Then the root mean square fluctuation, with different scale variable k of after detrended by is calculated by:(3)
In the final step, the logarithm of is plotted as a function of the logarithm of the time scale k. If the time series is self-similar, the scale-free property of a fractal geometry, this plot will display a linear scaling region, and the slope of the plot, alpha = , is called the scaling or self-similarity coefficient. If alpha = 0.5, is uncorrelated white noise, if alpha = 1.5, is Brownian noise, and if alpha = 1, is a 1/f power-law process widely existed in the real world [21].

With the obtained alpha from DFA, we defined the parameter a in equation (1) as a = -c/alpha, where c is a constant. In order to ensure the pitch vary from 0 to 96 in the range of the 128 pitch steps in MIDI (Musical Instrument Digital Interface), we chose c = 40 [12]. For the data of Subjects A and B, there are twoscale-free regions for each subject (Figure 2) [12], [21], [22].

thumbnail

Figure 2. EEG detrended fluctuation analysis and Scale-free regions.

(a) Subject A, the left region alpha = 1.022, and the right region alpha = 0.195; (b) Subject B, the left region alpha = 0.878, and the right region alpha = 0.215. Here k is the window size, and F(k) is the fluctuation from the local trends in windows.

doi:10.1371/journal.pone.0049773.g002

3.2 Intensity

1) Defined with EEG AP.

In our previous work [12], the intensity of a music note(MI) was proportional to the logarithm of the AP change according to the Fechner's law [11].(5)
Where l and k are two constants. In this approach, MI was partly related with Pitch, since both of them are defined with something related to the amplitude of EEG.

2) Defined with fMRI signal.

In this work, we proposed to adopt fMRI signal instead of the AP change of EEG to represent intensity of music. And as fMRI signal follows power law (Figure 3) [15], the resulted intensity of EEG-fMRI music would follow the power law too. Furthmore, as the sample rate of fMRI (about 2 seconds) is much lower than the usual music tempo, an interpolation step is necessary. In order to keep the power law and the underneath fractal structure, we adopted a fractal interpolation algorithm [23] to increase the sample rate. Figure 4 illustrates the interpolation results of the data used in this work. The new sample rate of the interpolated series is 1 second, which is close to the rate of a peaceful music.

thumbnail

Figure 3. fMRI signal detrended fluctuation analysis and Scale-free regions.

(a) Subject A, alpha = 1.013; (b) Subject B, alpha = 0.875. Here k is the window size, and F(k) is the fluctuation from the local trends in windows.

doi:10.1371/journal.pone.0049773.g003
thumbnail

Figure 4. Fractal Interpolation of the fMRI signals.

Points indicate the original sample, the solid lines indicate the interpolated function.(a) and (b) for Subject A and B, respectively.

doi:10.1371/journal.pone.0049773.g004

3.3 The mapping rules

The mapping rules between the brain physiological signals and the attributes of a music note are shown in Figure 5, where the fMRI signal reflected the BOLD signal, the EEG reflected the neural electrical activities. To denote the difference, we take the new music as EEG-fMRI music, whose intensity is fMRI based, and the previous one as EEG music, its intensity is EEG AP change based [12].

thumbnail

Figure 5. Composition rules of EEG-fMRI music.

The amplitude, waveform period, and fMRI signal are mapped to pitch, duration and intensity, respectively. The mappings from amplitude to pitch and from fMRI signal to intensity are based the power law.

doi:10.1371/journal.pone.0049773.g005

Results

1 EEG music

With the rules defined in paper [12], the EEG music (Figure 6(a)–(b)) was obtained from the EEG data (Figure 1). The EEG music (Audio S1, S2) sounds reasonably. However, as the pitch and intensity are derived from the amplitude and the induced AP, they were correlated significantly (Figure 6(c)–(d)).

thumbnail

Figure 6. EEG music.

Each column denotes a note, where the pitch (height of a column) is determined by Equation (1), the duration (width of a column) is defined as the period of an EEG wave, and the intensity (color of a column) is determined by Equation (5). (a) EEG music of Subject A with b1 = 96, b2 = 108, = 1.02, = 0.20; (b) EEG music of Subject B with b1 = 96, b2 = 108, = 0.88, = 0.21; (c) the relation between the pitch and intensity of the EEG music of Subject A with correlation coefficient = 0.427 (p<0.05); (d) the relation between the pitch and intensity of the EEG music of Subject B with correlation coefficient = 0.494(p<0.05).

doi:10.1371/journal.pone.0049773.g006

2 EEG-fMRI music

With the newly defined translation rules (Figure 5), the EEG-fMRI music (Figure 7, Audio S3, S4)was obtained from the EEG-fMRI data(Figure 1). In these music, as the pitch and intensity were defined separately by EEG amplitude and fMRI signal, they were not correlated directly. In fact, the correlation coefficient between pitch and intensity of the EEG-fMRI music is smaller than 0.01 (p>0.05) (Figure 7), which is much smaller than the case in EEG music (Figure 6), and this phenomenon is similar to a general man-made music. Figure 8 shows two pieces of score of EEG-fMRI music of the two subjects, and as music score only records the pitch and duration, the scores of EEG-fMRI music are the same with the EEG music, and the difference between them can only be recognized by listening to the Audios (See Audio S1, S2, S3, S4).

thumbnail

Figure 7. EEG-fMRI music.

Each column denotes a note, where the pitch (height of a column) is determined by Equation (1), the duration (width of a column) is defined as the period of the EEG wave, and the intensity (color of a column) is determined by fMRI signal. (a) EEG-fMRI music of Subject A; (b) EEG-fMRI music of Subject B. The parameters adopted here are the same with Figure 6; (c) The relation between the pitch and intensity of the EEG-fMRI music of Subject A with correlation coefficient = 0.0048 (p>0.05); (d) The relation between the pitch and intensity of the EEG-fMRI music of Subject B with correlation coefficient = 0.0095(p>0.05).

doi:10.1371/journal.pone.0049773.g007
thumbnail

Figure 8. Illustration of the score of EEG-fMRI music (Printed by Sibelius 4.0).

(a) Score of brain music of Subject A, (b) score of brain music of Subject B.

doi:10.1371/journal.pone.0049773.g008

3 Difference between EEG and EEG-fMRI music

In order to evaluate the difference between EEG music and EEG-fMRI music, 10 persons who have received music training for at least 3 years were invited to listen to the two subjects' music. In the test, 5 of them listened to the EEG music first and the other 5 listened to EEG-fMRI music first. For the question that which intensity change was quicker, 8(9) of them chose EEG music of Subject A(B), and for the question that which intensity change was slower, 9(9) of them chose EEG-fMRI music of subject A(B) (Table 1). The average identification of EEG music is 85%, and the average identification of EEG-fMRI music is 90%. Based on t-test, the overall evaluation is significant (EEG Music, T = 2.30, P<0.05; EEG-fMRI Music, T = 4.017, P<0.05).

thumbnail

Table 1. Intensity Variation recognition by 10 volunteers.

doi:10.1371/journal.pone.0049773.t001

These results indicate that the two kinds of music are quite different in intensity movement. And they all reported that the intensity change speed of the EEG-fMRI music was more close to usual human made scores.

4 Power law of the EEG and EEG-fMRI music

4.1 Power law of the EEG music.

Based on the translation rule (Equation.(1)) and the scale-free property of the EEG amplitude data (Figure 2), the pitch of EEG music obeys the power law rule [12].For the intensity and duration of EEG music, Figure 9 shows their DFA results, respectively, they clearly indicate that the scale index of each case (alpha) of each case is much smaller than 0.5, thus they all belong to “uncorrelated white noise”. This fact means that for EEG music only pitch has imitated the usual man-made music.

thumbnail

Figure 9. Detrended fluctuation analysis of EEG-music duration and volume.

(a)–(b) DFA of EEG music note duration for subjects A and B, respectively. For Subject A, alpha = 0.109, for Subject B, alpha = 0.056. (c)–(d) DFA of EEG music note volume for subjects A and B, respectively. For Subject A, alpha = 0.123, for Subject B, alpha = 0.082. Here k is the window size, and F(k) is the fluctuation from the local trends in windows.

doi:10.1371/journal.pone.0049773.g009
4.2 Power law of the EEG-fMRI music.

According to Figure 3, the fMRI signal, which represents the intensity of the EEG-fMRI music, obeys the power law rule. Therefore, from EEG music to EEG-fMRI music, not just the pitch but also the volume of music obeys the power law rule.

Discussions and Conclusions

1 Intensity and pitch

For EEG music [12], due to the quick change of the EEG state, the intensity of EEG music changed quickly and abruptly (Figure 6, Audio S1, S2), and this is not the usual case in man-made music. To reduce the gap, in this work, we chose another brain information, the fMRI signal, serving as the intensity information source. As the EEG-fMRI intensity evolution is smooth and leisure,the resulted EEG-fMRI sounds more close to the man-made real music (Figure 7, Audio S3, S4).

Besides, pitch and intensity, as two important factors in music composing, should be independent to each other in general. However, in the above EEG music, the amplitude of EEG is translated to pitch, and the AP change of EEG was used to represent the intensity of music, thereafter they two are not properly separated with each other. While in the above new EEG-fMRI music, the fMRI BOLD signal was used to represent intensity, it is almost independent to the EEG amplitude, untied the unreal close relation between pitch and intensity in EEG music thus better fit the usual composing regulation.

About the future of the EEG-fMRI music, the fMRI signal needs to be further carefully selected. As different frequency band of the fMRI signal may have different functional role, other than the adopted 0.01–0.08 Hz fMRI can be evaluated

2 Scale free music of the brain

Either EEG or fMRI is a physiologic signal. If we translate them to music directly, the physiological information sounds completely insert into the ‘music’, however, sucha ‘music’ may be just like noise. On the other hand, if we just use some EEG feature to trigger a man-made music piece, the music might be very pleasurable, but the physiological information was less involved. Therefore, a valuable and reasonable method would be a trade-off of these two extremes. In our approaches, both the previous EEG music and this new EEG-fMRI music, we assume the translation should follow some common rules obeyed by both brain signal and music. We believe that this is the correct way to understand human body through auditory. In our work, we pay special attention on scale-free phenomena, and as the scale-free phenomena exist widely in nature, including music and neural activities, it would be a reflection of the underneath truth of the mental state of the brain. The EEG-fMRI makes a step ahead the previous EEG music by extending the pitch scale-free of EEG music to both pitch and intensity scale-free of EEG-fMRI music, thus it provides us a new window to look inside the brain.

3 Spatio-temporal music

FMRI, which reflects the brain inherent metabolic activities, has a high spatial resolution,from which the music generated may display more details in spatial than any other currently available physiologic signal. Nevertheless, the EEG, collected on the surface of brain, is of high temporal resolution, could reflect brain's instantaneous activities. Therefore, it would be an interesting topic to develop a fusion method which combine EEG and fMRI to get the specific EEG-fMRI activities inside the brain [24], [25]. Then it would be more reasonable to make a vivo spatio-temporal music with EEG and fMRI which was derived from the same location inside the brain in the future.

4 Physiological music and man-made music

Table 1 reveals the distinct difference between the EEG and EEG-fMRI music, and the all subjects also reported that the intensity change speed of the EEG-fMRI music was more close to the usual human made scores. Here in order to display the intensity characteristics, we recorded the two types of music of Subject A by SONAR 6.0 and measured the variations of the envelope of the music waveforms, which represented the intensity changes of music. As contrast, we also measured a piece of real music ‘Nocturnes’, composed by Mozart. The results are shown in Figure 10. It is clear that the variation ranges of the intensity of the EEG-fMRI music waveform's envelope are similar with that of Mozart's music. Both of them are below 5 db. However, the changes of the intensity of EEG music are much wider. This fact means that the EEG-fMRI music translation method is better in mimic the real music. In addition, we could argue that the EEG-fMRI music do may better reflect the physiological brain process as the experiment state of Subject A was on resting state, and the intensity of her EEG-fMRI music does change slightly, just like a really peaceful music ‘Nocturnes’.

thumbnail

Figure 10. The envelope variation ranges of music waveform.

(a) A piece of real music ‘Nocturnes’, composed by Mozart, (b) EEG-fMRI music and (c) EEG music.

doi:10.1371/journal.pone.0049773.g010

5 Conclusion

In this work, we proposed a new method to translate both brain EEG and fMRI signals to music to better reflect the internal functional activities of the brain under the power law framework. The resulted music sounds better in mimic the man-made music intensity change. The brain music, as one of the human brain's intelligence product, embodies the secret of brain in an artistic style, provides the platform for scientist and artist to work together to understand ourselves, and it is also a new interactive link between the human brain and music. We hope the on-going progresses of the brain signals based music will properly unravel part of the truth in the brain, and then to be used for clinical diagnosis and bio-feedback therapy in the future.

Supporting Information

Audio S1.

30 s EEG music of Subject A from the resting state.

doi:10.1371/journal.pone.0049773.s001

(MP3)

Audio S2.

30 s EEG music of Subject B from the resting state.

doi:10.1371/journal.pone.0049773.s002

(MP3)

Audio S3.

30 s EEG-fMRI music of Subject A from the resting state.

doi:10.1371/journal.pone.0049773.s003

(MP3)

Audio S4.

30 s EEG-fMRI music of Subject B from the resting state.

doi:10.1371/journal.pone.0049773.s004

(MP3)

Acknowledgments

Thank Professor Dong Zhou and Qiyong Gong in West China Hospital for helping us collect the EEG-fMRI data in West China Hospital, Sichuan Province, China. Thank Zhao Gao for revising the language of this paper.

Author Contributions

Conceived and designed the experiments: DY C. Li JL. Performed the experiments: JL C. Luo. Analyzed the data: JL DW HY. Contributed reagents/materials/analysis tools: JL DW HY. Wrote the paper: JL DY.

References

  1. 1. Patel AD (2010) Music, language, and the brain: Oxford University Press
  2. 2. Bucci W (2001) Pathways of Emotional Communication. Psychoanalytic Inquiry: A Topical Journal for Mental Health Professionals 40–70. doi: 10.1080/07351692109348923
  3. 3. Adrian ED, Matthews BC (1934) The Berger rhythm: Potential changes from the occipital lopes in man Brain. A Journal of Neurology 57: 355–385. doi: 10.1093/brain/57.4.355
  4. 4. Miranda ER, Brouse A (2005) Interfacing the brain directly with musical systems: on developing systems for making music with brain signals. Leonardo 38: 331–336. doi: 10.1162/0024094054762133
  5. 5. Sousa AS, Baquero F, Nombela C (2005) The making of thegenoma music. Revista Iberoamericana de Micologia 22: 242–248. doi: 10.1016/s1130-1406(05)70050-4
  6. 6. Dunn J, Clark MA (1999) Life music: the sonification of proteins. Leonardo 32: 25–32. doi: 10.1162/002409499552966
  7. 7. Arslan B, Brouse A, Castet J, Filatriau J, Lehembre R, et al.. (2005) Biologically-driven musical instrument. eNTERFACE '05.
  8. 8. Rosenboom D (1990) Extended musical interface with the human nervous system International Society for the Arts,. Sciences and Technology doi: 10.1196/annals.1284.037
  9. 9. Wu D, Li C, Yin Y, Zhou C, Yao D (2010) Music Composition from the Brain Signal: Representing the Mental State by Music. Computational Intelligence and Neuroscience 267671. doi: 10.1155/2010/267671
  10. 10. Baier G, Hermann T, Stephani U (2007) Event-based sonification of EEG rhythms in real time. ClinNeurophysiol 118: 1377–1386. doi: 10.1016/j.clinph.2007.01.025
  11. 11. Teich MC, Heneghan C, Lowen SB, Ozaki T, Kaplan E (1997) Fractal character of the neural spike train in the visual system of the cat. J Opt Soc Am A 14: 529–546. doi: 10.1364/josaa.14.000529
  12. 12. Wu D, Li C, Yao D (2009) Scale-Free Music of the Brain. PloS ONE 4: e5915. doi: 10.1371/journal.pone.0005915
  13. 13. Richard FV, John C (1975) ‘1/f noise’ in music and speech. Nature 258: 317–318. doi: 10.1038/258317a0
  14. 14. Logothetis NK, Pauls J, Augath M, Tinath T, Oeltermann A (2001) Neurophysiological investigation of the basis of the fMRI signal. Nature 412: 150–157. doi: 10.1038/35084005
  15. 15. Biyu JH, John MZ, Abraham ZS, Marcus ER (2010) The Temporal Structures and Functional Significance of Scale-free Brain Activity. Neuron 66: 353–369. doi: 10.1016/j.neuron.2010.04.020
  16. 16. Qin Y, Xu P, Yao D (2010) A comparative study of different references for EEG default mode network: The use of the infinity reference. Clinical neurophysiology 121: 1981–1991. doi: 10.1016/j.clinph.2010.03.056
  17. 17. Yao D (2001) A method to standardize a reference of scalp EEG recordings to a point at infinity. Physiological Measurement 22: 693–711. doi: 10.1088/0967-3334/22/4/305
  18. 18. Collins DL, Zijdenbos AP, Kollokian V, Sled JG, Kabani NJ, et al. (1998) Design and construction of a realistic digital brain phantom. IEEE Transactions on Medical Imaging 17: 463–468. doi: 10.1109/42.712135
  19. 19. Fang G, Xia Y, Lai Y, You Z, Yao D (2010) Long-range corelations of different EEG derivations in rats: sleep stage-dependent generators may play a key role. Physiol Meas 31: 795–808. doi: 10.1088/0967-3334/31/6/005
  20. 20. Peng CK, Havlin S, Stanley HE, Goldberger AL (1995) Quantification of scaling exponents and crossover phenomena in nonstationary heartbeat time series. Chaos 5: 82–87. doi: 10.1063/1.166141
  21. 21. Gao T, Wu D, Huang Y, Yao D (2007) Detrended Fluctuation Analysis of the Human EEG during Listening to Emotional Music. J Elect Sci Tech 5: 272–277.
  22. 22. Hwa RC, Ferree TC (2002) Scaling properties of fluctuations in the human electroencephalogram. Phys Rev E 66: 021901. doi: 10.1103/physreve.66.021901
  23. 23. Penn AI (1997) Estimating fractal dimension with fractal interpolation function models. Medical Imging IEEE Transactions on 6: 930–937. doi: 10.1109/42.650889
  24. 24. Lei X, Qiu C, Xu P, Yao D (2010) A parallel framework for simultaneous EEG/fMRI analysis: Methodology and simulation. NeuroImage 52: 1123–1134. doi: 10.1016/j.neuroimage.2010.01.024
  25. 25. Yao D (2000) Electric potential produced by a dipole in a homogeneous conducting sphere. IEEE Trans Biomed Eng 47: 964–966. doi: 10.1109/10.846691