In the past years, a few methods have been developed to translate human EEG to music. In 2009, PloS One 4 e5915, we developed a method to generate scale-free brainwave music where the amplitude of EEG was translated to music pitch according to the power law followed by both of them, the period of an EEG waveform is translated directly to the duration of a note, and the logarithm of the average power change of EEG is translated to music intensity according to the Fechner's law. In this work, we proposed to adopt simultaneously-recorded fMRI signal to control the intensity of the EEG music, thus an EEG-fMRI music is generated by combining two different and simultaneous brain signals. And most importantly, this approach further realized power law for music intensity as fMRI signal follows it. Thus the EEG-fMRI music makes a step ahead in reflecting the physiological process of the scale-free brain.
Citation:Lu J, Wu D, Yang H, Luo C, Li C, et al. (2012) Scale-Free Brain-Wave Music from Simultaneously EEG and fMRI Recordings. PLoS ONE 7(11): e49773. doi:10.1371/journal.pone.0049773
Editor: Yong He, Beijing Normal University, Beijing, China
Received: July 30, 2012; Accepted: October 12, 2012; Published: November 14, 2012
Copyright: © 2012 Lu et al. This is an open-access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Funding:The authors acknowledge the following funding supports: the 863 projects 2012AA011601, 2012BAI16B02, the Natural Science Foundations of China No. 90820301, 60835005, 81201159, 60736029 and the ‘111’ project (B12027). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
Music and language define us as human . Emotional expression and communication, through language or non-linguistic artistic expression, are recognized as being strongly linked to health and sense of well-being . Therefore, as an artistic expression, music may represent human mind or mood.
In 1934, Adrian and Matthews attempted to listen to the brainwave signals via an amplified speaker . During the past decades, scientists and artists found many methods to make an electroencephalogram(EEG) sonification, although it is difficult for composition to balance music principles and EEG features . Meanwhile, in order to learn more about ourselves, researchers also used the deoxyribonucleic acid(DNA) , proteins , electromyograms (EMGs)  to compose music in the last century.
From the 1990s, various new EEG music generating rules were created . One of them was to translate some parameters of EEG to the parameters of music , and another one was to utilize some characteristics such as the epileptic discharges to trigger specific music , or to link brain states to various music pieces through Brain Computer Interface .
In 2009, we proposed a method to translate EEG to music. The translation rules included the direct mapping from the period of an EEG waveform to the duration of a note, the logarithmic mapping of the average power (AP) change of EEG to music intensity according to the Fechner's law , and a scale-free based mapping from the amplitude of EEG to music pitch according to the power law . However, in this method, the pitch and intensity were not independent enough under the translation rules as both pitch and intensity are related to EEG amplitude, so that the music was not strictly in accordance with the composition regulation in which pitch and intensity are usually not mutually related. Meanwhile, the intensity of music usually follows power law , however, the intensity of our previous brainwave music was obtained from the AP change of EEG within a time window, it didn't obey the power law (Shown in the following Figure 9). In this work, in order to imitate the general music composition better, we selected another brain information to represent the intensity instead of the EEG amplitude. As the intrinsic metabolic functional activities based functional magnetic resonance imaging(fMRI) is widely used to study the operational organization of the human brain , and fortunately, the fMRI blood oxygenation level dependent(BOLD) signal does follow the power law , thus the currently widely adopted fMRI may provide us a potential information for intensity of brainwave music. In fact, the fMRI BOLD signal is indirectly related to the electrical activities of a group of neurons by neuro-vascular coupling relation, thus it may reflect the brain mental state.
Data and Methods
1 Ethics Statement
This study was approved by the Research Ethics Board at University of Electronic Science and Technology of China. All participants were asked to read and sign an informed consent form before participating in the study. After experiment, all participants received monetary compensation for their time and effort.
2 EEG-fMRI Data
For the simultaneous EEG-fMRI recordings, the subjects were a 31-year-old female (Subject A) and a 14-year-old female (Subject B). They were both in resting state and scanned in a 3T MRI scanner (EXCITE, GE Milwaukee, USA).
For composing music, the EEG recordings were re-referenced to zero with a software called REST developed in our laboratory , . In this work, we chose the EEG at Cz electrode for brainwave music composition, which is at the central of the head and is a channel less affected by the body movement, and took the fMRI signal at the MNI(Montreal Neurological Institute)  coordinate (15,−48, 60), which was just below the electrode position Cz. In this way, we assumed the EEG and fMRI signals were almost from the same neural mass. The signals used in this work are illustrated in Figure 1.
(a) and (d) the fMRI signals collected at MNI coordinate (15,−48, 60); (b) and (e) pieces of signals in (a) and (d) for the following music compositions; (c) and (f) 30 s simultaneous EEGs collected at scalp Cz, respectively for Subjects A and B.
3 EEG-fMRI brain Music
Music note consists of four characters, pitch, timbre, duration and intensity. In this work, we paid special attention to pitch and intensity. And timbre was fixed with piano, which could be changed according to person's hobbies,while the duration was determined by the period of an EEG waveform.
In this study, we still adopt the power law rule between the amplitude (Amp) of an EEG waveform and the Pitch of a musical note ,(1)In equation (1), b is the maximum value of all pitches. Parameter a denotes the scale characteristics, and it is determined by the following detrended fluctuation analysis (DFA) , .
For a discrete time series , the first step of DFA is to subtract the mean from the series and then create a new series by integration:(2)where denotes the mean of the time series. Next, seriesis divided into a number of segments with length k (k represents the time scale of observation). For each of these segments, a local least-squares linear fit is conducted, and the resulted piece wise linear fit function is designated. Then the root mean square fluctuation, with different scale variable k of after detrended by is calculated by:(3)In the final step, the logarithm of is plotted as a function of the logarithm of the time scale k. If the time series is self-similar, the scale-free property of a fractal geometry, this plot will display a linear scaling region, and the slope of the plot, alpha = , is called the scaling or self-similarity coefficient. If alpha = 0.5, is uncorrelated white noise, if alpha = 1.5, is Brownian noise, and if alpha = 1, is a 1/f power-law process widely existed in the real world .
With the obtained alpha from DFA, we defined the parameter a in equation (1) as a = -c/alpha, where c is a constant. In order to ensure the pitch vary from 0 to 96 in the range of the 128 pitch steps in MIDI (Musical Instrument Digital Interface), we chose c = 40 . For the data of Subjects A and B, there are twoscale-free regions for each subject (Figure 2) , , .
(a) Subject A, the left region alpha = 1.022, and the right region alpha = 0.195; (b) Subject B, the left region alpha = 0.878, and the right region alpha = 0.215. Here k is the window size, and F(k) is the fluctuation from the local trends in windows.
1) Defined with EEG AP.
In our previous work , the intensity of a music note(MI) was proportional to the logarithm of the AP change according to the Fechner's law .(5)Where l and k are two constants. In this approach, MI was partly related with Pitch, since both of them are defined with something related to the amplitude of EEG.
2) Defined with fMRI signal.
In this work, we proposed to adopt fMRI signal instead of the AP change of EEG to represent intensity of music. And as fMRI signal follows power law (Figure 3) , the resulted intensity of EEG-fMRI music would follow the power law too. Furthmore, as the sample rate of fMRI (about 2 seconds) is much lower than the usual music tempo, an interpolation step is necessary. In order to keep the power law and the underneath fractal structure, we adopted a fractal interpolation algorithm  to increase the sample rate. Figure 4 illustrates the interpolation results of the data used in this work. The new sample rate of the interpolated series is 1 second, which is close to the rate of a peaceful music.
(a) Subject A, alpha = 1.013; (b) Subject B, alpha = 0.875. Here k is the window size, and F(k) is the fluctuation from the local trends in windows.
Points indicate the original sample, the solid lines indicate the interpolated function.(a) and (b) for Subject A and B, respectively.
3.3 The mapping rules
The mapping rules between the brain physiological signals and the attributes of a music note are shown in Figure 5, where the fMRI signal reflected the BOLD signal, the EEG reflected the neural electrical activities. To denote the difference, we take the new music as EEG-fMRI music, whose intensity is fMRI based, and the previous one as EEG music, its intensity is EEG AP change based .
1 EEG music
With the rules defined in paper , the EEG music (Figure 6(a)–(b)) was obtained from the EEG data (Figure 1). The EEG music (Audio S1, S2) sounds reasonably. However, as the pitch and intensity are derived from the amplitude and the induced AP, they were correlated significantly (Figure 6(c)–(d)).
Each column denotes a note, where the pitch (height of a column) is determined by Equation (1), the duration (width of a column) is defined as the period of an EEG wave, and the intensity (color of a column) is determined by Equation (5). (a) EEG music of Subject A with b1 = 96, b2 = 108, = 1.02, = 0.20; (b) EEG music of Subject B with b1 = 96, b2 = 108, = 0.88, = 0.21; (c) the relation between the pitch and intensity of the EEG music of Subject A with correlation coefficient = 0.427 (p<0.05); (d) the relation between the pitch and intensity of the EEG music of Subject B with correlation coefficient = 0.494(p<0.05).
2 EEG-fMRI music
With the newly defined translation rules (Figure 5), the EEG-fMRI music (Figure 7, Audio S3, S4)was obtained from the EEG-fMRI data(Figure 1). In these music, as the pitch and intensity were defined separately by EEG amplitude and fMRI signal, they were not correlated directly. In fact, the correlation coefficient between pitch and intensity of the EEG-fMRI music is smaller than 0.01 (p>0.05) (Figure 7), which is much smaller than the case in EEG music (Figure 6), and this phenomenon is similar to a general man-made music. Figure 8 shows two pieces of score of EEG-fMRI music of the two subjects, and as music score only records the pitch and duration, the scores of EEG-fMRI music are the same with the EEG music, and the difference between them can only be recognized by listening to the Audios (See Audio S1, S2, S3, S4).
Each column denotes a note, where the pitch (height of a column) is determined by Equation (1), the duration (width of a column) is defined as the period of the EEG wave, and the intensity (color of a column) is determined by fMRI signal. (a) EEG-fMRI music of Subject A; (b) EEG-fMRI music of Subject B. The parameters adopted here are the same with Figure 6; (c) The relation between the pitch and intensity of the EEG-fMRI music of Subject A with correlation coefficient = 0.0048 (p>0.05); (d) The relation between the pitch and intensity of the EEG-fMRI music of Subject B with correlation coefficient = 0.0095(p>0.05).
(a) Score of brain music of Subject A, (b) score of brain music of Subject B.
3 Difference between EEG and EEG-fMRI music
In order to evaluate the difference between EEG music and EEG-fMRI music, 10 persons who have received music training for at least 3 years were invited to listen to the two subjects' music. In the test, 5 of them listened to the EEG music first and the other 5 listened to EEG-fMRI music first. For the question that which intensity change was quicker, 8(9) of them chose EEG music of Subject A(B), and for the question that which intensity change was slower, 9(9) of them chose EEG-fMRI music of subject A(B) (Table 1). The average identification of EEG music is 85%, and the average identification of EEG-fMRI music is 90%. Based on t-test, the overall evaluation is significant (EEG Music, T = 2.30, P<0.05; EEG-fMRI Music, T = 4.017, P<0.05).
These results indicate that the two kinds of music are quite different in intensity movement. And they all reported that the intensity change speed of the EEG-fMRI music was more close to usual human made scores.
4 Power law of the EEG and EEG-fMRI music
4.1 Power law of the EEG music.
Based on the translation rule (Equation.(1)) and the scale-free property of the EEG amplitude data (Figure 2), the pitch of EEG music obeys the power law rule .For the intensity and duration of EEG music, Figure 9 shows their DFA results, respectively, they clearly indicate that the scale index of each case (alpha) of each case is much smaller than 0.5, thus they all belong to “uncorrelated white noise”. This fact means that for EEG music only pitch has imitated the usual man-made music.
(a)–(b) DFA of EEG music note duration for subjects A and B, respectively. For Subject A, alpha = 0.109, for Subject B, alpha = 0.056. (c)–(d) DFA of EEG music note volume for subjects A and B, respectively. For Subject A, alpha = 0.123, for Subject B, alpha = 0.082. Here k is the window size, and F(k) is the fluctuation from the local trends in windows.
4.2 Power law of the EEG-fMRI music.
According to Figure 3, the fMRI signal, which represents the intensity of the EEG-fMRI music, obeys the power law rule. Therefore, from EEG music to EEG-fMRI music, not just the pitch but also the volume of music obeys the power law rule.
Discussions and Conclusions
1 Intensity and pitch
For EEG music , due to the quick change of the EEG state, the intensity of EEG music changed quickly and abruptly (Figure 6, Audio S1, S2), and this is not the usual case in man-made music. To reduce the gap, in this work, we chose another brain information, the fMRI signal, serving as the intensity information source. As the EEG-fMRI intensity evolution is smooth and leisure,the resulted EEG-fMRI sounds more close to the man-made real music (Figure 7, Audio S3, S4).
Besides, pitch and intensity, as two important factors in music composing, should be independent to each other in general. However, in the above EEG music, the amplitude of EEG is translated to pitch, and the AP change of EEG was used to represent the intensity of music, thereafter they two are not properly separated with each other. While in the above new EEG-fMRI music, the fMRI BOLD signal was used to represent intensity, it is almost independent to the EEG amplitude, untied the unreal close relation between pitch and intensity in EEG music thus better fit the usual composing regulation.
About the future of the EEG-fMRI music, the fMRI signal needs to be further carefully selected. As different frequency band of the fMRI signal may have different functional role, other than the adopted 0.01–0.08 Hz fMRI can be evaluated
2 Scale free music of the brain
Either EEG or fMRI is a physiologic signal. If we translate them to music directly, the physiological information sounds completely insert into the ‘music’, however, sucha ‘music’ may be just like noise. On the other hand, if we just use some EEG feature to trigger a man-made music piece, the music might be very pleasurable, but the physiological information was less involved. Therefore, a valuable and reasonable method would be a trade-off of these two extremes. In our approaches, both the previous EEG music and this new EEG-fMRI music, we assume the translation should follow some common rules obeyed by both brain signal and music. We believe that this is the correct way to understand human body through auditory. In our work, we pay special attention on scale-free phenomena, and as the scale-free phenomena exist widely in nature, including music and neural activities, it would be a reflection of the underneath truth of the mental state of the brain. The EEG-fMRI makes a step ahead the previous EEG music by extending the pitch scale-free of EEG music to both pitch and intensity scale-free of EEG-fMRI music, thus it provides us a new window to look inside the brain.
3 Spatio-temporal music
FMRI, which reflects the brain inherent metabolic activities, has a high spatial resolution,from which the music generated may display more details in spatial than any other currently available physiologic signal. Nevertheless, the EEG, collected on the surface of brain, is of high temporal resolution, could reflect brain's instantaneous activities. Therefore, it would be an interesting topic to develop a fusion method which combine EEG and fMRI to get the specific EEG-fMRI activities inside the brain , . Then it would be more reasonable to make a vivo spatio-temporal music with EEG and fMRI which was derived from the same location inside the brain in the future.
4 Physiological music and man-made music
Table 1 reveals the distinct difference between the EEG and EEG-fMRI music, and the all subjects also reported that the intensity change speed of the EEG-fMRI music was more close to the usual human made scores. Here in order to display the intensity characteristics, we recorded the two types of music of Subject A by SONAR 6.0 and measured the variations of the envelope of the music waveforms, which represented the intensity changes of music. As contrast, we also measured a piece of real music ‘Nocturnes’, composed by Mozart. The results are shown in Figure 10. It is clear that the variation ranges of the intensity of the EEG-fMRI music waveform's envelope are similar with that of Mozart's music. Both of them are below 5 db. However, the changes of the intensity of EEG music are much wider. This fact means that the EEG-fMRI music translation method is better in mimic the real music. In addition, we could argue that the EEG-fMRI music do may better reflect the physiological brain process as the experiment state of Subject A was on resting state, and the intensity of her EEG-fMRI music does change slightly, just like a really peaceful music ‘Nocturnes’.
(a) A piece of real music ‘Nocturnes’, composed by Mozart, (b) EEG-fMRI music and (c) EEG music.
In this work, we proposed a new method to translate both brain EEG and fMRI signals to music to better reflect the internal functional activities of the brain under the power law framework. The resulted music sounds better in mimic the man-made music intensity change. The brain music, as one of the human brain's intelligence product, embodies the secret of brain in an artistic style, provides the platform for scientist and artist to work together to understand ourselves, and it is also a new interactive link between the human brain and music. We hope the on-going progresses of the brain signals based music will properly unravel part of the truth in the brain, and then to be used for clinical diagnosis and bio-feedback therapy in the future.
30 s EEG music of Subject A from the resting state.
30 s EEG music of Subject B from the resting state.
30 s EEG-fMRI music of Subject A from the resting state.
30 s EEG-fMRI music of Subject B from the resting state.
Thank Professor Dong Zhou and Qiyong Gong in West China Hospital for helping us collect the EEG-fMRI data in West China Hospital, Sichuan Province, China. Thank Zhao Gao for revising the language of this paper.
Conceived and designed the experiments: DY C. Li JL. Performed the experiments: JL C. Luo. Analyzed the data: JL DW HY. Contributed reagents/materials/analysis tools: JL DW HY. Wrote the paper: JL DY.
- 1. Patel AD (2010) Music, language, and the brain: Oxford University Press
- 2. Bucci W (2001) Pathways of Emotional Communication. Psychoanalytic Inquiry: A Topical Journal for Mental Health Professionals 40–70. doi: 10.1080/07351692109348923
- 3. Adrian ED, Matthews BC (1934) The Berger rhythm: Potential changes from the occipital lopes in man Brain. A Journal of Neurology 57: 355–385. doi: 10.1093/brain/57.4.355
- 4. Miranda ER, Brouse A (2005) Interfacing the brain directly with musical systems: on developing systems for making music with brain signals. Leonardo 38: 331–336. doi: 10.1162/0024094054762133
- 5. Sousa AS, Baquero F, Nombela C (2005) The making of thegenoma music. Revista Iberoamericana de Micologia 22: 242–248. doi: 10.1016/s1130-1406(05)70050-4
- 6. Dunn J, Clark MA (1999) Life music: the sonification of proteins. Leonardo 32: 25–32. doi: 10.1162/002409499552966
- 7. Arslan B, Brouse A, Castet J, Filatriau J, Lehembre R, et al.. (2005) Biologically-driven musical instrument. eNTERFACE '05.
- 8. Rosenboom D (1990) Extended musical interface with the human nervous system International Society for the Arts,. Sciences and Technology doi: 10.1196/annals.1284.037
- 9. Wu D, Li C, Yin Y, Zhou C, Yao D (2010) Music Composition from the Brain Signal: Representing the Mental State by Music. Computational Intelligence and Neuroscience 267671. doi: 10.1155/2010/267671
- 10. Baier G, Hermann T, Stephani U (2007) Event-based sonification of EEG rhythms in real time. ClinNeurophysiol 118: 1377–1386. doi: 10.1016/j.clinph.2007.01.025
- 11. Teich MC, Heneghan C, Lowen SB, Ozaki T, Kaplan E (1997) Fractal character of the neural spike train in the visual system of the cat. J Opt Soc Am A 14: 529–546. doi: 10.1364/josaa.14.000529
- 12. Wu D, Li C, Yao D (2009) Scale-Free Music of the Brain. PloS ONE 4: e5915. doi: 10.1371/journal.pone.0005915
- 13. Richard FV, John C (1975) ‘1/f noise’ in music and speech. Nature 258: 317–318. doi: 10.1038/258317a0
- 14. Logothetis NK, Pauls J, Augath M, Tinath T, Oeltermann A (2001) Neurophysiological investigation of the basis of the fMRI signal. Nature 412: 150–157. doi: 10.1038/35084005
- 15. Biyu JH, John MZ, Abraham ZS, Marcus ER (2010) The Temporal Structures and Functional Significance of Scale-free Brain Activity. Neuron 66: 353–369. doi: 10.1016/j.neuron.2010.04.020
- 16. Qin Y, Xu P, Yao D (2010) A comparative study of different references for EEG default mode network: The use of the infinity reference. Clinical neurophysiology 121: 1981–1991. doi: 10.1016/j.clinph.2010.03.056
- 17. Yao D (2001) A method to standardize a reference of scalp EEG recordings to a point at infinity. Physiological Measurement 22: 693–711. doi: 10.1088/0967-3334/22/4/305
- 18. Collins DL, Zijdenbos AP, Kollokian V, Sled JG, Kabani NJ, et al. (1998) Design and construction of a realistic digital brain phantom. IEEE Transactions on Medical Imaging 17: 463–468. doi: 10.1109/42.712135
- 19. Fang G, Xia Y, Lai Y, You Z, Yao D (2010) Long-range corelations of different EEG derivations in rats: sleep stage-dependent generators may play a key role. Physiol Meas 31: 795–808. doi: 10.1088/0967-3334/31/6/005
- 20. Peng CK, Havlin S, Stanley HE, Goldberger AL (1995) Quantification of scaling exponents and crossover phenomena in nonstationary heartbeat time series. Chaos 5: 82–87. doi: 10.1063/1.166141
- 21. Gao T, Wu D, Huang Y, Yao D (2007) Detrended Fluctuation Analysis of the Human EEG during Listening to Emotional Music. J Elect Sci Tech 5: 272–277.
- 22. Hwa RC, Ferree TC (2002) Scaling properties of fluctuations in the human electroencephalogram. Phys Rev E 66: 021901. doi: 10.1103/physreve.66.021901
- 23. Penn AI (1997) Estimating fractal dimension with fractal interpolation function models. Medical Imging IEEE Transactions on 6: 930–937. doi: 10.1109/42.650889
- 24. Lei X, Qiu C, Xu P, Yao D (2010) A parallel framework for simultaneous EEG/fMRI analysis: Methodology and simulation. NeuroImage 52: 1123–1134. doi: 10.1016/j.neuroimage.2010.01.024
- 25. Yao D (2000) Electric potential produced by a dipole in a homogeneous conducting sphere. IEEE Trans Biomed Eng 47: 964–966. doi: 10.1109/10.846691