Dissociation of the N 400 component between linguistic and non-linguistic processing : A source analysis study

The N400 component is commonly associated with the detection of linguistic incongruity. A few studies have shown that the N400 can also be elicited by non-linguistic stimuli. Different spatiotemporal patterns were observed between the typical Linguistic N400 and the Non-linguistic N400, suggesting distinct brain generators. The aim of this study was to investigate the presence of an N400 in response to linguistic and non-linguistic stimuli, and to specify anatomical sources of both N400s using a novel analysis method: the Bayesian Model Averaging (BMA) distributed source model. Picture-word and environmental soundpicture associations, either congruent or incongruent, were presented to ten young healthy adults while highdensity ERP recordings were made. Standard electrophysiological analyses confirmed that the N400 was not specific to linguistic incongruity but was also elicited by environmental sound-picture incongruities. Different topographic distributions were obtained for the Linguistic N400 and Non-linguistic N400. BMA analysis showed that the Linguistic N400 generators were mostly located in the left superior temporal gyrus, whereas the sources of the Non-linguistic N400 were identified mostly in the right middle and superior temporal gyri. Detection of linguistic incongruities recruited cerebral areas commonly associated with language processing, whereas non-linguistic incongruities recruited right cerebral regions usually associated with auditory processing of non-linguistic stimuli. The Linguistic and Non-linguistic N400s appear to be elicited by similar cognitive mechanisms assumed by different cerebral areas depending on the type of material to be processed. The present findings support the existence of parallel pathways for the processing of linguistic and non-linguistic incongruities.


INTRODUCTION
Event-related potentials (ERPs) provide an online measure of cortical activity [1] and have been useful in studying cognitive processes.Some ERP components, such as the N400, can provide specific information associated with semantic processing.The N400 is a negative deflection, which peaks between 350 and 600 ms after stimulus onset [2][3][4].This component was originally observed when subjects read sentences that ended with semantically incongruent words (for example: "I take my coffee with cream and engine") [5,6].The N400 would then reflect the participant's reprocessing of the unusual sentence ending.Kutas and Hillyard [7] further showed that the degree of unpredictability, and not only that of semantic incongruity, could increase the amplitude of the N400 (for example: "The bill was due at the end of the hour").
The N400 was also elicited in tasks involving the presentation of word pairs presented either visually or aurally [8,9], word-picture pairs [10], and word-iconic gesture pairs [11].For instance, Connolly et al. [10] using a computer-based adaptation of the Peabody Picture Vocabulary Test-Revised (PPVT-R) [12], presented a picture and an auditory word stimulus with a 700 ms onset asynchrony.Half of the picture-word pairs were correctly associated (e.g. the auditory word stimulus "ball" presented with the picture of a ball), whereas the other half of pairs were incorrectly associated (e.g. the auditory word stimulus "car" presented with the picture of a ball).Incorrectly associated pairs elicited significantly larger N400 amplitudes than correctly associated pairs.A French adaptation of this task was developed for the present study.
Source localization (or source reconstruction) enables the investigation of the neural generators of ERP components.Localization of these generators involves addressing the EEG Inverse Problem.This consists in the estimation of the distribution of the Primary Current Density (PCD) inside the brain that produces a given measured EEG.There are infinite solutions to this problem and thus specific additional prior information (or constraints) about the EEG generators must be included in order to obtain a unique and physiologically valid solution.Depending on the type of prior information used, there is a whole spectrum of different methods available.
Some methods handle the many-to-one nature of this problem (many parameters to estimate and very few data available) by characterizing the sources in terms of a limited number of current dipoles (dipolar model) that are fitted to the data through the minimization of some measure of the reconstruction error [35][36][37][38].In such models, each dipole represents the summed activity of a small cortical region and assumes that these dipoles vary only in strength over time [39].Over the last fifteen years, a number of studies attempted to localize the generators of the Linguistic N400 using different models for source localization.A summary of the results of these studies is reported in Table 1.For example, using a dipolar model approach, three MEG studies identified a left postero-superior temporal lobe source for the Linguistic N400 [40][41][42] while a fourth MEG study [43] reported bilateral auditory cortex generators for the same component.EEG studies using a dipolar model localized the sources of the Linguistic N400 along the left perisylvian cortex [44,45], bilateral anterior medial temporal lobe [46] or in bilateral superior temporal gyrus [47], while Silva-Pereyra et al. [48] reported a very wide distribution of Linguistic N400 sources in bilateral inferior, middle, and superior temporal regions, inferior frontal and insular areas, anterior cingulate as well as occipito-temporal and occipital areas.The growing experimental evidence about the existence of more diffuse brain networks has led to the emergence of a second method for source localization: the so-called distributed inverse solutions (DIS).These solutions address the non-uniqueness of the EEG inverse problem by using both anatomical and mathematical constraints.The anatomical constraints assume that the EEG sources are located in those brain regions capable of generating voltage fields over the scalp surface (gray matter for example), while the mathematical constraints make physiologically based assumptions about the mathematical properties of the solution obtained (like maximum spatial smoothness as in LORETA, for more details see [60] Trujillo- Barreto et al., 2004).
Several studies reported in Table 1 used a DIS approach to localize the sources of the Linguistic N400.For instance, Halgren et al. [52] visually presented sentences ending with semantically congruous or incongruous words to healthy adults during MEG recording.They found a Linguistic N400 occurring at around 250 ms that originated from the left Wernicke's area, lying in or near the superior temporal sulcus.The N400 activation then spread at about 270 ms to the anterior temporal sites and later on, at about 300 ms, to Broca's area and, finally, at about 370 msec, to the anterior orbital and fronto-polar cortices.In another study, Schulz et al. [58] who used LORETA and LAURA source localization algorithms with EEG data, reported that the generators of the Linguistic N400 elicited by their sentence-reading task were predominantly located in the left anterior middle temporal area.Similarly, Giglio et al. [50] and Travis et al. [59] found a left fronto-temporal N400 source distribution using distributed source models.However, Khateb et al. [53] investigated, using LAURA, source localization of N400s elicited from different linguistic tasks including different stimulus types (words vs. images) and showed the involvement of the bilateral middle/superior temporal gyrus in all tasks.Hamm et al. [47], who used the LORETA source localization algorithm, reported right frontal and bilateral temporal generators, whereas Haan et al. [51] found a very broad and scattered distribution for the generators of the Linguistic N400, which they attributed to important inter-individual differences impeding accurate localization of the sources.
To sum up, investigation of the Linguistic and Nonlinguistic N400 components revealed differences in their spatial distribution.For the Linguistic N400, findings of the studies reported in Table 1 show a tendency to localize more commonly sources of the Linguistic N400 in the left temporal lobe.Nonetheless, the results regarding the localization of the generators of the Linguistic N400 from these studies show some inconsistencies.The discrepancy may stem, either from differences in the tasks that were used to elicit the Linguistic N400, some of these tasks included linguistic and non-linguistic stimuli within the same experimental paradigm, or from differences in the models and algorithms that were used for source localization.Development of more reliable and adapted DIS for late component such N400 is thus needed.
To our knowledge, no electrophysiological studies have so far been conducted in order to localize the cerebral generators of the Non-linguistic N400 elicited by environmental sounds.The only source of information comes from a study by Kotz et al. [61] who reported that a N400 was elicited by meaningful environmental sounds in healthy adults but not in patients with either left or right anterior temporal lobe lesion.The present set of experiments aims to clarify whether or not the Linguistic and the Non-linguistic N400 are distinct components generated by different groups of neurons.Among the non-linguistic sounds, environmental sounds are those that received the most attention.Environmental sounds are construed as the best non-linguistic stimuli homologous to human speech because they have a complex spectral structure, and they are easily identifiable [62].Results of the studies conducted with brain-damaged patients [63] as well as with normal subjects using fMRI [64] or PET [62,65,66] suggested that environmental sounds are mostly processed in the right hemisphere.Environmental sounds are thus well-suited stimuli to use in a task eliciting a Non-linguistic N400, because the generators of this component are likely to be located in the right hemisphere, in contrast with the Linguistic N400 whose generators are generally located in the left hemisphere [40,42,44,52].
The present study constitutes the first attempt to investigate anatomical sources of both the Linguistic N400 and Non-linguistic N400 (elicited by an environmental sound-picture task).To this end, picture-auditory word and environmental sound-picture pairs that were erroneously or correctly matched were presented to participants during ERP recording.Source analyses based on a distributed Bayesian model averaging (BMA) approach [60] were performed on both N400 components.
The BMA approach is based on the application of the Bayesian model inference scheme [67] using evidence approximation [68], to the EEG/MEG Inverse Problem.It has been used recently by Olivares and colleagues [19] to localize a N400-like effect elicited by a familiar faces matching task.The Bayesian paradigm allows estimating the posterior probability of each model given the data, which represents a measure of "how good" that model is for explaining the data.This measure of goodness expresses a trade off between goodness-of-fit (data reconstruction error based on that model) and complexity (number of parameters used by the model to explain the data) of the model.The number of parameters of the given model is proportional to the total number of voxels contained in the anatomical regions used to constrain the inverse solution in that model.That is, higher probabilities are assigned to the simpler models that explain the data better.Then, the model uncertainty is taken into account by averaging the primary current densities (PCDs) obtained for each particular anatomical constraint (or model) weighted by their model posterior probabilities.This weighted average provides a PCD, which is unconditional on any of the anatomical constraints assumed.In this approach, the solution under each anatomical constraint is obtained using the traditional LORETA method [69].

Participants
The sample was composed of 10 young healthy adults (five males, age range = 20 -28 years with a mean age of 24.78 years).All participants were native speakers of Canadian French.The data from one male participant had to be rejected because of electrophysiological artefacts.Informed consent was obtained from all participants prior to their inclusion in the study.The project was approved by the Ethics Committees of the Sainte-Justine and Notre-Dame University of Montreal Hospital Centers.Moreover, all participants gave their informed consent prior to their inclusion in the study.

Tasks and Stimuli
All participants performed two tasks 1) the environmental sound-picture matching task, which consisted of determining whether a presented picture was congruent or not to an environmental sound (non-linguistic task) and 2) the picture-auditory word matching task, consisting in determining whether an auditory word was congruent or not to a picture (linguistic task).

Environmental Sounds
The 180 environmental sounds (for example, bird songs, car noises, waterfall sounds, etc.) used in this study were taken from the Sounds Effects Library created by Lucas-Film Ltd & Sound Ideas.Onset and offset of each sound was determined using Cool Edit Pro TM software allowing the creation of sounds that lasted exactly 2383 ms.Environmental sounds were presented with Optimus 40232 model XTS24 speakers at 75 dB.Ninety environmental sounds were correctly associated to a picture (e.g. the noise of a truck paired with the picture of a truck) and 90 others were incorrectly paired to the same set of pictures (e.g. a laugh, paired with the picture of a truck).

Words
All 180 words were digitally recorded in stereo in a male voice using Cool Edit Pro TM software.Word duration varied from 327 to 1104 ms, with a mean duration of 684 ms.Onset and offset of each word were determined visually with PRAAT TM software based on the word's digitized waveform, allowing precise cutting of each word.Words were presented with Optimus 40232 model XTS24 speakers at 75 dB.The 90 auditory words correctly associated with a picture were taken from the Échelle de Vocabulaire en Images Peabody (EVIP) [70].The 90 words that were erroneously matched to the same set of pictures were selected from different word lists [71][72][73][74].They were paired with the set of congruent words based on acquisition age, number of syllables, lexical category, and the phonetic features of the first phoneme, which was either a voiced or unvoiced stop (e.g.b, t) or voiced or unvoiced fricative (e.g.v, f).In the incongruent pairs, the auditory word and the picture did not belong to the same semantic category nor did they share phonological similarity.For instance, the picture of a boat (the French word for boat is "bateau") was paired with the heard word "gorille" (meaning "gorilla" in English).

Pictures
All 180 pictures paired with auditory stimuli (environmental sounds and words) were simple black and white drawings of 420 × 315 pixels (a visual angle of approximately 7 degrees) taken from the EVIP, the French version of the PPVT-R, and from various web sites.Pictures derived from EVIP A and B booklets were selected among items from lowest performance levels and were electronically digitized with ScanJet 5300C HP TM and PrecisionScan TM 3.03 software.All pictures were resized to a 1024 × 768 pixels resolution using Adobe Photoshop software and were presented in the middle of a ViewSonic Professional Series P225F screen (refreshing rate of 75 Hz), which was placed 85 cm from the subject.Timing and sequence of pictures were controlled by a computer using MatLab 6.5 software and auditory stimuli were presented by a Tucker-Davis System (RP2.1;Real-time Processor).In each task, all pictures were presented twice because the same set of pictures was used for both congruous and incongruous pairs.Pilot studies were performed to ensure that both tasks were successfully performed by participants (success rate over 90%) and induced a N400 effect.These studies in-dicated that the N400 was best elicited with the sound preceding the picture in the environmental sound-picture matching task and the opposite was found for the picture-auditory word matching task.Figure 1(a) illustrates the procedure used in the environmental sound-picture matching task.Ninety pictures were paired with a congruent and an incongruent environmental sound.Each trial consisted of an environmental sound, which appeared 1400 ms before the picture onset and remained audible during picture presentation and 483 ms after the offset of the picture.The duration of the picture presentation was 500 ms.A trigger was placed at the beginning of the presentation of each picture and was used to time-lock EEG sampling to picture onset.
In the picture-auditory word matching task (see Figure 1(b)), another set of 90 pictures was paired with a congruent and an incongruent word.Each trial consisted of a picture, which appeared 700 ms before the onset of the auditory word and remained visible during the auditory word presentation (mean duration was 684 ms) and 1000 ms after the offset of the word.A trigger was placed at the beginning of each word and was used to time-lock EEG sampling to word onset.
The Linguistic N400 is elicited by auditory stimuli (picture-auditory word matching task) whereas the Non-linguistic N400 is evoked by pictorial material (environmental sound-picture matching task).In previous studies, both pictorial stimuli and auditory stimuli have been found to elicit similar N400-like effects in response to semantic mismatch [15].Nevertheless, a paired Hotelling T2 test (incongruent versus congruent) was applied on the data of each task in order to eliminate common and other early components when performing source analyses (see extended description in the Source Analyses Section).

Procedure
Participants performed both the picture-auditory word matching and environmental sound-picture matching tasks in dim light while sitting in a comfortable chair in a , which appeared 1400 ms before picture onset.Every picture appeared on the screen for a duration of 500 ms (red rectangle).The sound remained on for the entire trial for a total trial duration of 2383 ms.(a2) Each of the 90 pictures was correctly or erroneously paired with a sound.For example, for a congruent trial, the sound of a "boat" would be paired with a picture of a "boat" and for an incongruent trial the sound of a "cat" would be paired with the same picture of a "boat"; (b) Picture-auditory word matching task.(b1) The picture (red rectangle) appeared 700 ms before the word onset and remained visible during the entire trial (mean total trial duration of 2384 ms).Word mean duration was 684 ms (blue rectangle).(b2) Each of the 90 pictures was correctly or erroneously associated to a word.sound-proof, electrically shielded room.Both tasks were composed of 4 practice trials followed by 180 experimental trials, and were administered during electrophysiological recording (see "Electrophysiological recording" section).The order of trial presentation was pseudo-randomized for both tasks; all trials were presented randomly with the exception of trials with the same picture, which were separated by at least one other trial.A 2-second inter-trial interval consisted of a blank screen with a central fixation cross ("+").Tasks were divided in four blocks of 45 trials, for a total of 180 experimental trials for each task.Since each trial had a duration of 4383 ms, each block had a total duration of approximately 197 seconds (approximately 3 minutes 29 seconds).Short periods of rest (1 -2 minutes) were given to participants between blocks to reduce fatigue and eye strain.All participants performed both tasks twice in order to increase the total number of trials from 180 to 360 in each task, for a total duration of approximately 60 minutes including inter-block and inter-task periods of rest.The total testing duration did not exceed two and a half hours including electrode placing, experiment, and electrode location digitalization.During the placing of the electrodes and the subsequent digitalization, which both took approximately 60 minutes; the participant was quietly watching a movie.
During both tasks, the participant responded bimanually on a keyboard to each trial by pressing two green buttons when the auditory and visual stimulus association was congruent, and by pressing two red buttons when the association was incongruent.A bimanual response was selected in order to eliminate unilateral cortical activation that would have been associated with mono-manual responses.All participants were asked to avoid blinking while the sound was audible (environmental sound-picture matching task) or the picture was visible (picture-auditory word matching task), and to blink between trials, if needed.
The environmental sound-picture matching task was administered prior to the picture-auditory word matching task in order to prevent as much as possible, the subjects from verbalizing the visual and auditory stimuli of the environmental sound-picture matching task.Indeed, the picture-auditory word matching task involved heard words that automatically recruit language processing.Hence, the administration of the picture-auditory word matching task prior to the environmental sound-picture matching task would likely have accustomed and encouraged the participant to adopt a linguistic strategy in performing the subsequent environmental sound-picture matching task.

ERP Recordings and Analyses
EEG activity was recorded from 128 Ag/AgCl electrodes with a NeuroScan Synamps 2 TM system at a sampling rate of 500 Hz with a bandpass filter of 0.1 -100 Hz and a central reference.The electrodes were embedded in a 128 NeuroScan Quick-Cap TM .Impedances were maintained at or below 5 kΩ during the whole recording session and an electrode placed in the frontal area (between AFz and Fz) served as ground.Vertical and horizontal eye movements were recorded using external electrodes placed above, under, and on the outer canthus of both eyes.Heart rate and muscular activity were also recorded from, respectively, left and right electrodes placed on Erb point and on both sides of the neck.Evoked potentials were recorded with Neuroscan 4.3 software (Compumedics, USA).
Location of each electrode and three fiducial points (nasion, and left and right preauricular) was digitized and recorded with Brainsight TM Frameless system (Rogue Research, Montreal, Canada) on a Macintosh G4 computer to allow for source analysis.
EEG recordings were analysed using the Brain Vision Analyser program (Brain Products, Munich, Germany).A digital band pass filter of 0.1 -30 Hz and 24 dB/octave, as well as ocular correction [75] were applied off-line.Data were referenced to the averaged signal of all electrodes.The continuous EEG recordings were then epoched beginning 200 ms before, and ending 1000 ms after stimulus onset (601 data points).Segmented data were corrected using an artefact rejection (threshold of ±100 µv).Data were averaged across trials for each experimental condition after baseline correction using the 200-ms time segment before stimulus onset.
The N400 component was scored as the most negative peak occurring between 325 and 475 ms for the environmental sound-picture matching task and between 375 and 500 ms for the picture-auditory word matching task.N400 peak amplitudes were measured as the voltage difference between the mean activity for the 200-ms baseline period before stimulus onset and the most negative point within the specified latency range.N400 peak latencies corresponded to the time from stimulus onset and to the most negative point within the specified latency range.Topographical maps (Brain Vision) were computed from the normalized amplitude data [76,77].

Source Analyses
EEG data collected from each task (linguistic and nonlinguistic) and condition (incongruent and congruent) were processed separately.After averaging over all accepted trials, the specific N400 peak latency for each subject was identified as described in the previous section.Source analyses were performed individually at the N400 peak latency, on each task and condition separately using BMA approach.
An image of the underlying brain electrical activity (primary current density) for each individual scalp topography, for each task and condition, was estimated by BMA of different EEG inverse solutions (see [60] for details), each one based on anatomical constraints derived from the Montreal Neurological Institute MNI average-brain atlas.Individual lead fields were calculated using the Reciprocity Theorem [78], with a three spheres and isotropic conductivities head model [79,80].A constraint-independent final image of the underlying brain electrical activity related to each task and each condition was then obtained by averaging all computed primary current densities weighted by the corresponding support that they received from the data.This weighting coefficient was measured in terms of the posterior probability of the corresponding model (anatomical constraint), given the data.In order to measure statistical differences between both incongruent and congruent conditions of the currents estimated [(x, y, z) components] at each voxel, we used paired Hotelling T2 statistic tests for each task separately.Based on previous peak detection analyses (see section "ERP analyses"), the paired Hotelling T2 statistic test was applied in a time window between 325 and 475 ms for the non-linguistic task and between 375 and 500 ms for the linguistic task.The threshold for the resulting statistic parametric map was selected using false discovery rate [81] with q = 0.1.This test provided inverse solution statistical maps comparing both conditions for each task separately.Individual and Grand Average inverse solutions were visualized using the Brain Electrical Tomography (BET) viewer software (Neuronic S.A.), which integrates the functional information provided by the EEG with the anatomical information provided by the MRI.

Behavioral Analyses
Success rates for both tasks were measured to ascertain that the participants' attention level was acceptable and that they were performing the task adequately.Individual success rates varied between 91.11% and 99.72% for the environmental sound-picture matching task, and between 96.39% and 100% for the picture-auditory word matching task.On average, participants could accurately recognize 96.98% ± 3.22% of the congruent/incongruent stimulus pairs of the environmental sound-picture matching task and 98.73% ± 1.63% of congruent/incongruent stimulus pairs during the picture-auditory word matching task.No difference was found between congruent and incongruent trials.Although a Student t test for task difficulty difference did not reach significance (t 9 = 2.271, p = 0.053), the performance during the picture-auditory word matching task tended to be slightly higher than that observed for the environmental sound-picture matching task.Following testing, the participants were questioned to determine if they had used different strategies when performing the two tasks.They indicated that during the environmental sound-picture matching task, they tried to visualize the object corresponding to the sound they were hearing whereas during the picture-auditory word matching task, they silently anticipated the word corresponding to the picture being presented.

Standard Electrophysiological Analyses
Considering the high success rates obtained for both tasks, ERP analyses were conducted on all trials.Moreover, less than 10% of all trials were rejected due to artifact or ocular correction.Figure 2 shows the grand-average ERPs obtained when averaging the responses elicited by each stimulus for all participants during On average, the N400 elicited by the environmental sound-picture matching task (Figure 2(b)) began at around 298 ms post-stimulus onset.The largest N400 amplitude was found at FCz and had an amplitude of −3.603 µV with a latency of 360 ms for the incongruent condition and an amplitude of −2.638 µV with a latency of 352 ms for the congruent condition.A fronto-central topographical distribution was observed for the Non-linguistic N400 (Figure 2(b)).
The N400 component waveform found in the picture-auditory word matching task (Figure 2(a)) began on average at approximately 342 ms post-stimulus onset.The largest N400 was found in Cz where the amplitude was on average −1.485 µV with a latency of 414 ms for the incongruent condition and 0.097 µV with a latency of 416 ms for the congruent condition.The Linguistic N400 presented a centro-parietal distribution (Figure 2(a)).
Because previous work suggested that presentation of the same stimuli can reduce the amplitude and shorten the duration of the N400 (e.g.[82,83], split-half analyses on amplitude and latency data as well as on success rates were computed to address the issue of habituation.An ANOVA with time (1 versus 2) and condition (congruent versus incongruent) as factors revealed that N400 amplitudes and latencies did not differ between both administrations of the protocol (p > 0.1).Furthermore, performance did not improve with increased exposure to the task, as indicated by near zero statistical values for Time.Therefore, all 360 trials of both tasks were used in further analyses.

N400 Latency
The analysis of the N400 latencies was conducted on the ten midline electrodes because of its well-known centroparietal distribution of the N400 component (e.g.[5,33]).A three-way within-subject ANOVA with task (non-linguistic versus linguistic), condition (congruent versus incongruent), and region (average of the five most anterior midline electrodes covering frontal and fronto-central areas versus average of the five most posterior electrodes placed on centro-parietal and occipital regions) as factors was performed on the N400 latency data.This analysis revealed a main effect of task (F (1,8) = 41.70,p < 0.001, ε = 1.00) due to shorter Non-linguistic N400 latencies in response to pictures in the environmental sound-picture matching task (mean = 351.56± 27.70 ms) in comparison to Linguistic N400 latencies in response to auditory words presented in the picture-auditory word matching task (mean = 422.17± 33.15 ms).No interactions or other simple effects were found.
In order to further compare topographical data between both tasks, a t-test was conducted using BrainVision on normalized amplitude data [76,77].Given the Student's t-distribution table, it was assumed that in order for activity distribution to be significantly different across both tasks in the incongruent condition, the tvalue would have to be >3.36.The two N400 component distributions were spatially different (t 8 > 3.36, p < 0.05) between 340 and 430 ms: the Non-linguistic N400 (environmental sound-picture matching task) was associated with a fronto-central scalp distribution whereas the Linguistic N400 (picture-auditory word matching task) produced a centro-parietal distribution.This significant time window (340 -430 ms) has been further used for source analyses.
With the aim of investigating the presence of hemispheric asymmetries, a t-test was performed.Consistent with the literature, hemispheric asymmetries were found for amplitudes in the picture-auditory word matching task (t 8 = 2.788, p = 0.024), where Linguistic N400 amplitudes were larger on the right hemisphere (an average of the data from 20 electrodes overlying the right fron-to-centro-parietal area) than on the left hemisphere (an average of the data from 20 electrodes overlying the left fronto-centro-parietal area).No hemispheric asymmetries were found in the Non-linguistic N400 elicited by the environmental sound-picture matching task.

Electrophysiological Source Analyses
In order to determine if the N400 components elicited by the picture-auditory word and environmental sound-picture matching tasks had different N400 generator localizations, source analyses using a distributed BMA approach were performed on both N400s.Results suggested different cerebral sources for each component elicited by incongruous trials.In fact, the generators of the Non-linguistic N400 were situated in the right middle and superior temporal gyri, in bilateral parahippocampal gyri and in the left lingual gyrus (Figure 3(a)), whereas those for the Linguistic N400 were located in the left superior temporal gyrus and in the right parahippocampal gyrus (Figure 3(b)).Individual data showed that these results were found in approximately 70% of patients.

DISCUSSION
The aim of the present study was to further investigate the nature of a N400 elicited by incongruity detected in an auditory word-picture pair (Linguistic N400) stimulus and in an environmental sound-picture pair (Non-linguistic N400) stimulus, and to specify the anatomical sources of both N400s.As expected [5,6,22], our data showed a larger N400 in response to incongruent than to congruent stimulus pairs in both tasks, confirming that the N400 is not specific to the detection of lexico-semantic errors but can be elicited by the detection of a broader type of incongruities [26][27][28]30,84,85].Shorter peak latencies were found for the Non-linguistic N400, a finding that is consistent with other previous works [22,23,27,53].The observed differences in the N400 latency may stem from a quicker integration of the environmental sound and the corresponding picture due to the less arbitrary character of the relationship with its world-referent [13].
Alternatively, latency differences can be accounted for by the linguistic (or not) nature of the stimuli.Hence, conceptual processing of environmental sounds (ex: sound of yelp) could be faster than linguistic processing (ex: word "dog") because the access to word meaning is mediated through phonological representation whereas environmental sounds may activate more directly the corresponding semantic representations [22].
Consistent with the literature [5,10,33,54], the scalp topography of the Linguistic N400 showed a centro-parietal distribution with a right-greater-than-left hemispheric asymmetry.In contrast, the Non-linguistic N400 component showed a fronto-central scalp topography with no hemispheric asymmetry.Studies that included environmental sound-picture matching task showed controversial results since some reported a left-greater-than-right cerebral asymmetry [26,30,34] while others found no asymmetry [22,27,28] for the Non-linguistic N400.In the present study, no such hemispheric asymmetry was found.
Latency and topographical differences obtained between Linguistic and Non-linguistic N400s may reflect different processes relying on distinct cerebral generators, but may also be observed when the same cognitive operation is applied to process contexts (for example, images vs. sounds) that differ in their cortical representations [53].A way to avoid this limitation would have been to use written instead of auditory words in the linguistic task.However, since we wanted to develop a paradigm that is suitable for all populations, including young children and cognitively challenge individuals, we did not want the task to required reading abilities.Another way would have been to present the picture prior to environmental sound in the non-linguistic task.However, a pilot study showed that environmental sounds needed to have a significantly longer duration than words to be recognized and that duration greatly varies between sounds, which would have significantly influenced the non-linguistic N400 latency and its variability.Therefore, we cannot exclude that the cognitive process associated with each N400 is the same and that the observed differences are due to the cortical activation preceding/or occurring in parallel and generated by different types of stimuli.Nevertheless, a paired Hotelling T2 test was applied on the data of each task in order to eliminate early components when performing source analyses.
In our study, the picture-auditory word matching task was a French adaptation of the linguistic task used by Connolly et al. [10] that evoked a Linguistic N400 with a tendency for a right-greater-than-left cerebral asymmetry.This topographical pattern has commonly been reported in the literature [5,33], and was also found in the present study for the Linguistic N400.The results of the source analysis for the Linguistic N400 are also similar to those obtained in EEG [44], MEG [40,42,52,59] and fMRI [86] (for a review, see [87]), that all showed left superior temporal gyrus sources for the Linguistic N400.This cerebral region, located in Wernicke's area, is typically associated with phonological and semantic processing and a lesion in this area has been shown to induce comprehensive language deficits such Wernicke's aphasia.The localization of generators in this region is thus consistent with a N400 elicited by the detection of an incongruity occurring during auditory language processing.
Even if no hemispheric asymmetry was found in the Non-linguistic N400 topography, the results of the BMA source analysis suggested a laterality effect, in showing a greater involvement of right than left middle and superior temporal region in processing of the environmental sounds (see Figure 3).These results are compatible with results from brain imaging studies [62,64,66] showing that this region is involved in the processing of environmental sounds.Damage in this area may induce auditory agnosia [88], an auditory impairment affecting the processing of environmental sounds.
Results of the source analyses also revealed generators in the parahippocampal region that were common to both N400 components.The parahippocampal gyrus has been associated with recall activity [89].The contribution of this region may thus be related to the very nature of the tasks we used which both involved the access to representations that are stored in long-term memory.
In a recent study comparing N400 components elicited by linguistic tasks (semantic and phonologic) and a non-linguistic task (image categorization), Khateb et al. [53] found no topographical or sources analysis differ-ences between tasks.All N400s had a central distribution and originated from the middle and superior temporal gyri.In the present study, we found topographical differences between the Linguistic N400 (centro-parietal distribution) and Non-linguistic N400 (fronto-central distribution).Our source analyses suggested different brain generators for both Linguistic and Non-linguistic N400s, located respectively in the left superior temporal gyrus and in the right middle and superior temporal gyri.Differences between the present findings and results from Khateb et al. [53] may be explained by methodological differences.First, different tasks were used to elicit the Non-linguistic N400 components.It can be hypothesized that the environmental sound-picture matching task used in our experiment induces less verbalization than the image categorization task used in Khateb et al.' study [53].This could explain the fronto-central distribution associated with the Non-linguistic N400 and its right middle and superior temporal gyri origin found in our study.The fronto-central distribution could also be explained by a familiarity effect and/or the implication of the attention network.It is noteworthy that such a fronto-central distribution has also been reported in previous studies using environmental sound tasks [22,23,27,30].Second, differences in source analysis approaches may also partly explain differences between our results and those of Khateb et al. [53].
In the present study, we used a unique and powerful source analyses model, the BMA approach, including LORETA method.This is the first study investigating Linguistic N400 cerebral generators using the BMA approach.Most distributed inverse solutions (DIS) give quite good results when dealing with widespread activities.However, they fail to recover spatially concentrated sources due to their tendency to smooth out activations.The most common procedure followed by any DIS is, for any given mathematical constraint, to assume that the possible generators are confined to a given region of the brain and to make all statistical inferences conditional on that assumption.However, different choices of the anatomical constraints lead to completely different current density distributions.This introduces some uncertainty about the model assumptions, which must be taken into account.This problem, commonly omitted by traditional inverse solution methods, has been widely treated in the Bayesian Inference literature and it is known as Model Uncertainty.The BMA approach used in this study [60] is based on the application of the Bayesian model inference scheme [67] under the evidence approximation [68], to the EEG Inverse Problem.This provides a coherent mechanism for accounting for the type of model uncertainty described above.The BMA has shown significantly less blurring and ghost sources than alternative distributed inverse solution approaches and is significantly less affected by the underestimation of deep sources.
Data set collected in both tasks and conditions comprised a fairly high proportion of artifact-free trials.This was obtained by presenting each stimulus set twice.It is worth mentioning that we did not find the repetition effect reported by some authors [82,83,90,91] when measuring N400 elicited by the detection of incongruity in sentence stimuli.These authors showed that the N400 amplitude was reduced in the second presentation of the sentence stimulus set.According to these authors, the last word of a sentence gets more predictable at the second or multiple presentations.This predictability elicits a diminution of the N400 amplitude.In the present study, we used pictures, auditory words and environmental sounds.These stimuli may not provide a semantic context that is sufficiently rich to contribute in reducing the level of incongruity detected by the subjects at the second presentation.
Within the present set of experiments, it may be argued that the Non-linguistic N400 was elicited because the subjects verbalized the environmental sounds before the appearance of the picture, thereby eliciting a Linguistic N400.To address this issue, Aramaki et al. [92] only used sounds in their experimental design.However, this was not possible in the present study.In order to minimize this effect, all participants performed the environmental sound-picture matching task prior to the picture-auditory word matching task.In a pilot study with no EEG recording, we investigated the order effect on the performance of a small group of subjects and found that all participants obtained very high scores in both tasks (ceiling effect), with no task order effect.In addition, responses given by the participants questioned after the testing do not support the hypothesis of the verbalization during the environmental sound-picture matching task.Indeed, participants reported they were adopting different strategies when performing the picture-auditory word or the environmental sound-picture matching task, i.e., silently anticipating the word congruent with the picture prime in the picture-auditory word matching task, and visualizing the object or an environmental scene eliciting such a sound in the environmental sound-picture matching task.Thus, it can be hypothesized that the picture-auditory word matching task involved a verbalization of the name of the picture presented before the word onset, while the environmental sound-picture matching task induced a visual representation of the environmental sound presented before the picture onset.The fact that an activation of the visual associative cortex (left lingual gyrus) was found only in the environmental sound-picture task brings additional support to this interpretation.

CONCLUSION
The present paper is the first study that proposes models of cerebral generators for both the Non-linguistic N400, generated by an environmental sound-picture matching task, and the Linguistic N400, elicited by a pictureauditory word matching task, in healthy adults.Results for source analysis indicated that BMA approach is an adequate model for the localization of the N400.Findings of the present study concur with brain imaging data and single case studies suggesting that linguistic incongruities involve Wernicke's area, whereas those related to environmental sounds recruit the right temporal region.Our results are also consistent with those of a PET study conducted by Thierry and Price [65], who reported activation in the left superior temporal sulcus and left ventral inferior frontal gyrus in the processing of the linguistic incongruity, whereas the right midfusiform and right posterior middle temporal gyrus were activated during the processing of non-linguistic incongruity.Interestingly, hemispheric dissociation in processing linguistic and nonlinguistic material was obtained regardless of the modality of presentation (visual or auditory).It thus appears that the stimulus type, linguistic vs. non-linguistic, influences which cerebral regions will be recruited in detecting semantic incongruity.This suggests the existence of two parallel pathways for the processing of linguistic and non-linguistic incongruities.

Figure 1 .
Figure 1.Procedure used in both tasks.(a) Environmental sound-picture matching task.(a1) Each trial consisted of a sound (blue rectangle), which appeared 1400 ms before picture onset.Every picture appeared on the screen for a duration of 500 ms (red rectangle).The sound remained on for the entire trial for a total trial duration of 2383 ms.(a2) Each of the 90 pictures was correctly or erroneously paired with a sound.For example, for a congruent trial, the sound of a "boat" would be paired with a picture of a "boat" and for an incongruent trial the sound of a "cat" would be paired with the same picture of a "boat"; (b) Picture-auditory word matching task.(b1) The picture (red rectangle) appeared 700 ms before the word onset and remained visible during the entire trial (mean total trial duration of 2384 ms).Word mean duration was 684 ms (blue rectangle).(b2) Each of the 90 pictures was correctly or erroneously associated to a word.

Figure 2 (
a) picture-auditory word matching tasks and Figure 2(b) environmental sound-picture.

Figure 2 .
Figure 2. Grand-average ERPs obtained when averaging the responses elicited by each stimulus for all participants during both tasks.(a) Linguistic task.The largest Linguistic N400 was found in Cz where the amplitude was on average −1.485 µV with a latency of 414 ms for the incongrent condition (black line on the graph) and 0.097 µV with a latency of 416 ms for the congruent condition (red line on the graph).As shown on the right (a), the Linguistic N400 presented a centro-parietal distribution; (b) Non-linguistic N400 and its topography.The largest Non-linguistic N400 amplitude was found at FCz and had an amplitude of −3.603 µV with a latency of 360 ms for the incongruent condition (black line on the graph) and an amplitude of −2.638 µV with a latency of 352 ms for the congruent condition (red line on the graph).A fronto-central topographical distribution was observed for the Non-linguistic N400 (right b).In order to get a better display, only most relevant electrodes for both N400 are shown (FCz and Cz).

Figure 3 .
Figure 3. BMA source analyses results for (a) environmental sound-picture matching task and (b) picture-auditory word matching task.All individual inverse solutions have been average in order to visualize grand-average BMA source analyses.Results show (a, i) right (R) middle and superior temporal gyri (red arrow), bilateral parahippocampal gyri (yellow arrows) and left (L) lingual gyrus (white arrow) sources for the Nonlinguistic N400.Left ii) shows right (R) middle and superior temporal (auditory area) Non-linguistic N400 generators on a 3D brain reconstruction; (b) L superior temporal gyrus (red arrow) and R parahippocampal gyrus (yellow arrow) generators for the Linguistic N400.Right ii) shows clear L superior temporal (Wernicke's area) Linguistic N400 generators.N.B.As specified on the figure, the radiological convention is used: L = R; R = L.

Table 1 .
Linguistic N400 source analysis literature review (in alphabetic order) from dipole and distributed source models.Lt = left; Rt = right.