“Independent Researcher, Librarian, Music Educator and Composer" - Ted Hunter

Thought Reading Capacity


John J. McMurtrey, M. S., Copyright 2004,[a] 3 Jan 08 [b] 

Co-authorship is negotiable towards professional publication in an NLM indexed journal, Email- Johnmcmurt@aol.com 

Donations toward future research are gratefully appreciated at http://www.slavery.org.uk/FutureResearch.htm 

ABSTRACT 

            Electroencephalographic, Magnetoencephalographic, and functional Magnetic Resonance Imaging reports of specific concept recognition in humans on hearing words, viewing images or words, and prior to vocalization are examined.  These reports are consistent with an extensive literature on word category differentiation by electrophysiology and blood flow, which is reviewed.  EEG discrimination of emotional states, and deception literature is surveyed along with non-invasive brain computer interface reports.  Non-contact and remote methods of brain wave assessment are also considered.  The literature treated lends some substantiation to press accounts and case anecdotes that thought reading is possible, and has had covert development.

 

INTRODUCTION

The Bible attributes to God the capacity to know the thoughts of men. [1]  Most scientists are unaware that thought reading by electroencephalogram (EEG) was reported as feasible in work begun over 30 years ago, [2] which more recently a number of groups confirm by Electroencephalography (EEG), Magnetoencephalograpy (MEG), and functional Magnetic Resonance Imaging (fMRI) technologies.  This review focuses on literature relating to technologic thought reading, though also treated are the discrimination of more general cognitive states, brainwave capture methods, and reports of thought reading development, apparently covert to open literature.



METHODS OF SPECIFIC CONCEPT RECOGNITION

The Defense Advanced Research Projects Agency in 1972 contracted Pinneo & Hall for work that a 1975 US technical report entitled “Feasibility Study For Design of a Biocybernetic Communication System.”   The study concludes “that it is feasible to use the human EEG coincident with overt and covert speech as inputs to a computer for such communication” (covert speech is defined as verbal thinking). 2  The 149 page report [c] states:  “enough information has been obtained . . . to specify the optimum parameters to use for an EEG operating system, and to suggest future research towards that end.” 

Pinneo & Hall utilized templates for EEG word recognition constructed by averaging EEG patterns evoked by 9 words in each subject for visually presented words, and primarily utilized 4 electrodes over brain language areas for prediction.  People with high hemispheric lateralization had EEG patterns for some words that frequently classified 100% correctly, regardless of the number of repetitions with stability over time.  Over all words, however, classification accuracy for these people was 85% for overt speech, and 72% for words repeated to oneself, but solely by mental means without vocalization.  Across all subjects specific word EEG patterns were classified 35% correctly for overt speech, and 27% correctly for covertly spoken words, but more people were in the 70-100% classification range than in the 10-15% range. [d]  Subjects with low hemispheric laterality, particularly stutterers had near chance EEG classification.  EEG concept recognition was actually 10-15% higher for pictures rather than words.  Phrases containing similarly articulated words or homonyms were better recognized than these words alone without context.

A US Office of Naval Research funded government study reports characteristically distinct topography of amplitude and decay coefficient for sub-waveforms at frequencies from 15-60 Hz for recalled or viewed digits from 0 to 9, and the words yes or no at electrodes over left hemisphere Brodman’s Area 39/40 and right hemisphere occipital electrode position O2 in 3 individuals with some indication of similarities between subjects. [3]  Background activity unrelated to digit or word stimulus was cancelled out, and the author indicates that the usually studied Event Related Potential (ERP) waveform is a summed composite of sub-waveforms.  All other studies here discussed analyzed such a composite ERP.

Suppes et. al. have the most extensive recent publications supporting and reporting specific EEG thought recognition starting in 1997, a year before the above report. [4] [5] [6] [7] [8]  This work largely compares recognition improvement methods with some emphasis on a relative invariance of EEG concept representations across individuals.  The procedures generally utilized Fourier transforms of both templates for recognizing words, and test samples with an optimal EEG frequency window, or filter selected for each subject.   EEG word templates constructed by averaging each subject’s responses (50 trials) at single electrodes resulted in less EEG word recognition, 4 than recognition templates averaged across all subjects (700 trials) [e] for bipolar electrode difference. 6  The latter technique produced recognition rates over seven words of 100% for visual images and auditory words. 6 [f]  However, for visually presented words, recognition templates generated by excluding from the average the subject tested was better--75% than averaging within a subject or over all subjects.  The waveforms for each presentation modality were very similar, and when recognition templates averaged across subjects in the modalities of visual images or words were utilized for recognizing other modalities (visual images or words & by audition), recognition still was generally 60-75%.  Such results were despite inclusion of three subjects with English as a second language, and obvious hemispheric laterality confounds important to Pinneo & Hall, [g] such as one left handed and another ambidextrous subject.  These results indicate a relative invariance of EEG representations for different concepts between subjects and perception modality, when averaging out and filtering noise.  Matching templates to words is determined by the least amplitude difference between template and test word waveforms, when sampled at 814 difference points as squared and summed (Pinneo & Hall 2 had 255 samples per word).

Also examined are brain wave patterns for sentences.  Recognizing the first sentence word by the same words individually presented, and the same words in sentences when cut and pasted was successful at a 50% recognition rate (with 8.3% as chance). 5  However for differentiating whole sentences, over 90% recognition was obtained for 48 sentences, as visually presented one word at a time. 7 

Averaged unfiltered auditory responses are classified 100% correctly by the superposition of 3 sine waves chosen from the frequency domain maxima for each word.8  The same procedure when averaged across subjects and presentation modalities (visual images, visual and auditory words) classifies 100% of the words (or images) by 5 frequencies per concept, while data fit decreased only 6% compared to the filtered templates.  Syllable classification is less successful, with six correct classifications out of eight examples from superposition of nine frequencies.

Two subjects in Suppes et al. 1997 4 had 64 channel EEG recordings from which scalp current density can be calculated by the surface Laplacian, which filters artifacts from muscle activity.  Recognition rates could be improved by 9 % in one subject, and 4 % in the other. [9]  Both subjects had coincident foci maximally predicting recognition on the head.

Yes/no decision discrimination of 86% by spatio-temporal cross correlation is reported. [10]  This was achieved from 4 electrodes over bilateral frontal and occipital sites.  Differential equation measures of synchronization rate and average polarity also had high recognition rates of 78% and 81% respectively.

Magnetoencephalographic (MEG) recognition of viewed words is reported above chance significantly by 27% for recognition and 44% for accuracy [11] by a speech recognition classifier.  Suppes et al. 4 5 also investigated MEG word recognition with lesser results than for EEG, but there is reanalysis of some of this data by more advanced classifiers for words presented by audition, and viewed with instruction to silently say the word. [12]  Best single trial correct classification for heard words in a subject is 60.7 % over 9 words by Independent Components Analysis combined with Linear Discriminant Classification, but averages across subjects are 40.6 % for auditory, and 30.9 % for viewed words (words presented ranged from 7-12, so chance levels ranged from 8.3-14.3 %). 12  

There is apparently a Russian report of specific EEG word recognition before 1981. [13]  The work is only known from a science reporter, and specifically unavailable, but is mentioned to aid this report’s discovery, and because of the claim that specific words contain category information, which is of possible significance for word category differentiation studies.

Patents for EEG thought recognition exist.  Electroencephalographic (EEG) instant detection by syllables of “a content of category which the testee wishes to speak” quotes Kiyuna et. al. Patent # 5785653 “System and method for predicting internal condition of live body.” [14]  A stated use:  “the present invention may be use (sic) to detect the internal condition of surveillance in criminal investigation” by EEG.  NEC Corporation licensed this patent.  Mardirossian Patent # 6011991 “Communication system and method including brain wave analysis and/or use of brain activity” includes remote EEG communication with armed forces or clandestine applications. [15]   This patent proposes transponder capable skin implants, utilizes artificial intelligence, and is licensed by Technology Patents, LLC.

A classifier based on computational linguistics correctly identifies the functional Magnetic Resonance Imaging (fMRI) pattern for 77 % of 60 nouns as averaged over 9 subjects, as well as correct prediction of the fMRI pattern for 72 % of 1000 frequent words. [16]  Previous fMRI reports confirm similar capabilities for viewing pictures of objects with lesser classification methods.  Comparing the distributed brain activity observed by fMRI for viewing faces, houses, cats, chairs, bottles, shoes, and scissors were 90-100% correct in all two category comparisons (with 50% as chance). [17]  A different group confirms this analysis. [18]  Even though all these objects are described as categories because different exemplars and views were presented, discrimination of these objects generally requires an adjective, so that the distinctions qualify as specific concepts.  One report examined just 20 seconds of fMRI data rather than one half of an fMRI session in previous studies, and utilized different exemplars of an object category for training classifiers from those utilized during classification.  A support vector classifier provided the best results with 59-97% accuracy among ‘categories’ of baskets, birds, butterflies, chairs, teapots, cows, horses, tropical fish, garden gnomes, and African masks (with 10% as chance). [19]  “Brain reading” are descriptive terms titling this report.  Another study reports 78 % average correct classification (range 59-94 %) for viewing across all line drawing exemplars for a drill, hammer, screwdriver, pliers, saw, apartment, house, castle, igloo, or hut with even better discrimination when considered as categories of tools and dwellings. [20]  A quantitative fMRI receptive field model for the visual cortex could provide 92 and 72 % correct identification for 120 natural images novel to the viewing experience of each participant. [21]  Visual cortex response to 440 multi-grey scale checkerboard-like patterns is reported to train local decoders to reconstruct viewed images that are correctly identified among millions of candidate images, and that effectively reads out perceptual state. [22]  Though not here considered specific concepts, review of considerable ability to decode viewed line orientations or grids is available [23] that is expected related to fMRI and electrophysiological discrimination of viewed objects with one review considering such capacities as mind reading. [24]  Particularly remarkable of such studies is above chance discrimination of imagined specific patterns in some subjects considering the lesser brain activity, and classifier model [25] compared to [21].

Numerous fMRI studies show similarly activated brain regions for viewing images or words, and hearing words.  Viewing pictures of objects or the word naming them activates similar distributed brain systems for storing semantic knowledge, [26] [27] [28] and auditory presentation also shares the same [29] or a similar [30] system with that of viewing these words.  These studies give anatomical basis for the high cross modality recognition rates of concepts observed by Suppes et al. 6 8

 

PHYSIOLOGIC DISCRIMINATION OF WORD CATEGORIES

            Broca and Wernicke originally defined anatomy pertinent to aphasia resulting from brain injury. [31]  More recently described are brain lesion patients who have very selective agnosias, which is an inability to name or recognize specific object classes. [32] [33] [34]  Many word category differentiation reports reviewed below were initiated to explain and substantiate such deficits.  This literature is consistent with specific word recognition, because word responses are averaged by category, and distinguished with only statistical inspection without specific comparison to templates or by classifiers as is required for thought recognition.  Brain cell assembly activation provides a theoretical framework for both specific concept recognition, and word category discrimination. [35] 

 

Electroencephalographic and Magnetoencephalographic Word Category Discrimination

Evoked EEG responses discriminate nouns and verbs.  Nouns elicit more theta power than verbs, but verbs have greater theta coherence decrease, particularly in frontal versus posterior sites. [36]  Noun waveforms generally are more negative than verb responses at post-stimulus intervals of both 200-350 and 350-450 milliseconds (msec.). [37] [38] [39] [40]  Ambiguous noun/verbs are more negative than unambiguous nouns or verbs in the early latency interval, and when context indicates noun meaning versus verb use, are more negative over both these latency windows. 40  Anterior-posterior electrode activity also differs for ambiguous versus unambiguous nouns and verbs. 40 [41]  

Action verb waveforms differ in amplitude, 38 and central versus posterior distribution compared to visual nouns, [42] with particular 30 Hz increase over the motor cortex for action verbs, and over the visual cortex for visual nouns. [43] [44]  Face, arm, or leg action verbs differ in amplitude by time interval, and activity increases over the specific corresponding motor strip locus as well as by frontal electrode. [45] [46]  Low resolution electromagnetic tomography finds irregular verb activity more in the left superior and middle temporal gyri, while regular verbs are more active in the right medial frontal gyrus at 288-321 msec. [47]  Irregular verbs respond more in the left ventral occipito-temporal cortex than regular verbs at ~340 msec. by MEG, which localizes perpendicular sources undetectable by EEG. [48]  Regular verb activity modulates more the left inferior prefrontal region including Broca’s area at ~470 msec with MEG, but irregular verbs have more right dorsolateral prefrontal cortex activity at ~570 msec.  Priming evoked patterns occur for regular but not irregular verbs, [49] [50] while incorrect irregular noun plural [51] and verb participle [52] [53] waveforms differ from that of incorrect regular forms.

Abstract word waveforms onset more positively about 300 msec., persist longer at lateral frontal sites, and distribute more to both hemispheres compared to concrete words.38 [54] [55]  β-1 frequency coherence during memorization of concrete nouns indicates left hemisphere electrode T5 as the main brain processing node. [56]  Left hemisphere electrode T3 is similarly important for abstract nouns, which have more frontal area contribution, and massive right posterior hemisphere coupling. 56  Abstract versus concrete memorization distinctly changes other frequency bands, [57] [58] and theta synchronization predicts efficient encoding. [59]      

Content words yield a more negative peak at 350-400 msec. than functional grammar words, with a subsequent occipital positivity that function words lack, and more electrode and hemisphere differences from 400-700 msec. [60] [61]  In sentences, the late component of function words resembles preparatory slow waves that apparently subserve their introductory and conjunctive grammatical function. [62]  Other studies show content versus function word differences at additional intervals and more bi-hemispheric effects,[63] with right visual field advantage for function words. [64]  MEG distinguishes functional grammar words, or content words such as multimodal nouns, visual nouns, or action verbs, each by response strength and laterality at intervals of both ~100 and greater than 150 msec. [65] 

Proper name amplitudes peak more just after 100 msec. negatively, and just after 200 msec. positively than common nouns, while one’s own name accentuates these peaks relative to other proper names with further positive and negative components. [66]  Proper names, animals, verbs, and numerals show electrode site differences:  proper name temporal negativity extends to inferior electrodes bilaterally; verbs and animal names are less negative and similar, but verbs have left frontal inferior positivity; while numerals have less waveform negativity, and bilateral parietal positivity. [67]   Non-animal objects are more negative in both the 150-250 and 350-500 msec. intervals than animals, while animals are more positive in the 250-350 msec. interval. [68] [69]  Animals are more positive in approximately the same latter interval than vegetables/fruits, while vegetables/fruits are more negative in about the earlier interval (150-250 msec.), and have stronger frontal region current sources than animals. [70]  Animals in natural scenes evoke different waveforms than just natural scene or building pictures. [71]  Responses to words for living things are less negative over the right occipital-temporal region than artifactual objects, while pictorial presentations of the same items further differ and have hemisphere effects noted as unreported. [72]  EEG waveforms for specific meanings could be as discretely categorized as indicated by the reported but unspecified Russian work, which claims that “the waves for such concepts as “chair”, “desk”, and “table” are all overlapped by another wave that corresponds” to the concept of furniture. 13

Affective word meanings such as good-bad, strong-weak, or active-passive are discriminated [73] by both category and meaning polarity according to response latency, amplitude, and scalp distribution at intervals of 80-265 and 565-975 msec. [74]  Positive words have amplitude increases peaking at 230 msec. compared to negative words, and relative to neutral words increase a subsequent peak amplitude as well as a slow wave component. [75]  Emotional words also show less amplitude decrease on repetition than neutral words. [76]

Some of these word category differentiation reports are consistent with both the specific recognition reports, and/or the discrimination of non-verbal cognition.  Based on EEG/MEG responses, words are readily distinguished from non-words, [77] [78] [79] pictures, [80] and as to length. [81]  Even commas have a characteristic waveform similar to the speech phrase closure evoked pattern called closure positive shift. [82]  Color selection modulates the EEG. [83]  EEG discriminates the judgment of gender for both faces and hands. [84]

 

Positron Emission Tomography (PET) and Functional Magnetic Resonance Imaging (fMRI) Word Category Discrimination

Positron Emission Tomography (PET) and Functional Magnetic Resonance Imaging (fMRI) localize brain blood flow, with ability to distinguish perceptual categories.  Some studies locate recognition of places [85] [86] and faces [87] within certain brain areas, however, expertise can recruit the face recognition area, [88] and other studies show these areas only responding maximally for specific stimuli. [89]  Word category activity is both distributed and overlapping 89 [90] in a somewhat lumpy manner. [91]  Though regions of word category maximal difference are indicated below, brain comprehension is not solely dependent on these areas.  Discrete category responsive emergence may have some resemblance to category segregation in the feature processing of artificial neural networks that self organize without programming. [92]

Meta-analysis of 14 studies locating activity for face, natural, and manufactured object recognition shows ventral temporal cortex difference.  Face recognition activates more inferior ventral temporal portions including the fusiform gyrus of which manufactured objects activate more medial aspects than face or natural objects, yet natural objects distribute more widely in this region. [93]  Eighty eight percent of face studies converged for mid fusiform gyrus activity, while natural and manufactured objects converged no more than 50% for any discrete area.   Manufactured object activity locates to the middle temporal cortex from natural objects, which locate more in the superior temporal cortex.  Face and natural object activity is more bilateral, and in the left inferior frontal cortex, while particularly tools activate the premotor area.  These studies also feature activity in the inferior occipital/posterior fusiform as well as the medial occipital structures of lingual gyrus, calcarine sulcus, and cuneus.

There is some agreement that verbs have greater activity in temporal, parietal, and premotor/prefrontal regions than nouns, while nouns have little [94] or no [95] greater activated areas than verbs, yet no noun/verb difference is also reported. [96]  German regular noun and verb fMRI responses compared to irregular words differ significantly in the right precentral gyrus, the left prefrontal cortex, bilateral posterior temporal lobes, and bilateral complexes including superior parietal lobules, supramarginal gyri, and angular gyri. [97]  Regular words are left hemisphere lateralized, while irregular words have somewhat greater distribution to the right hemisphere, and a greater activation over all cortical areas.    Irregular verbs activate more total cortex than regular verbs, but lack motor strip, insular, and most occipital cortex activity present for regular verbs. [98]  Though both forms activate the inferior parietal lobule, irregular verbs activate more posterior and superior portions than regular verbs.

Depending on control task correction, naming actions activates the left inferior parietal lobule, which is lacking for locative prepositions, which activate the left supramarginal gyrus selectively from actions. [99]  Furthermore, naming abstract shape location compared to locating concrete items increases right supramarginal gyrus activity, 99 which specifically also activates on long-term memory for spatial relations  [100] and in American sign language prepositions. [101]  The supramarginal gyrus is encompassed by the temporal-parietal-occipital junction active for location judgments, and is separate from temporal activity for judging color. [102]  Action word generation activity is just anterior to the motion perception area, while color word generation activity is just anterior to the color perception area. [103]  Naming object color activates distinct brain regions from naming the object, with color knowledge retrieval activity being slightly removed from that of naming colors.  [104]  Irrespective of language and visual or auditory modality, the naming of body parts activates the left intraparietal sulcus, precentral sulcus, and medial frontal gyrus, while naming numbers activates the right post central sulcus as joined to the intraparietal sulcus. 29 

Concrete words are discriminated from abstract words in both noun or verb forms, 95 with more right hemisphere activity for abstract words than concrete words. [105] [106] [107]   Abstract/concrete contrasts feature both right or left temporal areas, while the reverse concrete/abstract comparison features frontal activity. [108] [109] [110] [111] [112]  Besides distinction from abstract nouns, the concrete categories of animals contrasted to implements respond selectively in the posterior-lateral temporal, and frontal cortex areas across studies. 105 110  Limbic activity, particularly the cingulate, distinguishes emotional words from both abstract and concrete words. 106     

Naming pictures of animals, tools, and famous people are discriminated [113] by increased regional blood flow in the left inferior frontal gyrus for animals, premotor area for tools, and left middle frontal gyrus for people. [114]  Faces activate the right lingual and bilateral fusiform gyri, while the left lateral anterior middle temporal gyrus response differs to famous faces, famous proper names, and common names. [115]  Particularly the left anterior temporal cortex responds to names, faces, and buildings when famous relative to non-famous stimuli. 115 [116]  Viewing photographs of faces, buildings, and chairs evokes activity distributed across several cortical areas, which are each locally different in the visual, ventral temporal 89 and occipital cortices. [117]  Photograph perception of these same categories has more hemispheric lateralization and activation than non-perceptual imagery, [118] while short term memory face imagery activity is stronger than that of long term memory. [119]  

More advanced fMRI techniques discriminate further word or object classes.  In a high resolution fMRI limited brain cross section study; the activity differs for animals, furniture, fruit, or tools in discrete sites of the left lateral frontal and 3 separate medial temporal cortex loci respectively. [120]  The application of artificial intelligence to fMRI patterns distinguishes between 12 noun categories (fish, four legged animals, trees, flowers, fruits, vegetables, family members, occupations, tools, kitchen items, dwellings, and building parts). [121]  Finally are the reports of discriminating the viewing  different ‘categories’ 17 18 19 so discrete as to require an adjective for distinction, and those acknowledging specific concept recognition 16 20 as well as prediction of photograph perception, 21 previously discussed.   

Some cognitive functions are related to or partly dependent on language.  Letters activate the left insula more than objects and exclusively activate the left inferior parietal cortex. [122]  Letters also activate an area in the left ventral visual cortex more than digits in most subjects. [123] [124]  Brain activations of mathematical thinking are partly dependent on language. [125]  Subtraction activates bilaterally the anterior intraparietal sulcus and a phoneme area in the intraparietal sulcus mesial to the angular gyrus, selectively from simple motor tasks. [126]  Number comparison activates right hemisphere intraparietal and prefrontal areas, while multiplication localizes more to the left hemisphere. [127]   

 

ELECTROENCEPHALOGRAPHIC DISCRIMINATION OF OTHER COGNITIVE STATES

Other literature indicates EEG differentiation of completely non-verbal cognition.  Greater left prefrontal activity predicts positive affect, while greater right prefrontal activity predicts negative disposition in psychological testing. [128]  However, the stability of hemispheric activation is important for such a trait characteristic, [129] and more transient mood states have exactly the opposite arousal symmetry. [130]   Decreased left prefrontal activity is also found in depression, [131] [132] and the anxiety situations of social phobics. [133]  Patented is more specific attitude, mood, and emotion differentiation, by plotting at least two and as many as five EEG frequencies, with reference to Air Force research. [134]   EEG patterns discriminate relative misanthropy and philanthropy in facial preferences, and favorable or negative responses to faces, [135] while waveform topography identifies sad face perception. [136]  Another EEG emotion indicator is the stimulus-preceding negativity (SPN).  Although slight SPNs can precede instruction cues, this wave is most pronounced while awaiting performance assessment and reward or aversive feedback. [137] [138] [139] [140]

A number of groups have developed procedures to detect deception based on the P300 (positive @ 300 millisec.) event related potential (ERP) from EEG. [141] [142] [143] [144] [145] [146]  Brain Fingerprinting is a commercial system, [147] which includes additional frequency analysis, particularly a late negative ERP potential, and cites 100% accuracy over five separate studies. [148] [149] [150] [151] [152]  Though most EEG deception detection concerns situation specific knowledge, a late positive potential approximate to the P300, is reported to vary as a function of real attitude rather than attitude report. [153] 

 

BRAIN COMPUTER INTERFACE 

EEG cortical potentials are detected for both actual movement, [154] and movement readiness potentials (bereitschaftspotential). [155] [156]  EEG sufficiently differentiates just the imagination of movement to operate switches, [157] move a cursor in one [158] or two dimensions, [159] control prosthesis grasp, [160] and guide wheel chairs left or right [161] for prompted responses.   EEG detects such potentials to play Pac Man, [162] and imagining the spinning of cubes, or arm raising in appropriate direction guides robots through simulated rooms, [163] [164] [165] both achieved without response prompting.  Unprompted slow cortical potentials also can turn on computer programs. [166]  Signals from implanted brain electrodes in monkeys achieve even more complex grasping and reaching robot arm control without body arm movement. [167]  Some ability to recognize evoked responses to numbers [168] and tones [169] in real time by a commercial system called BrainScope has limited report.

 

PROXIMATE AND REMOTE BRAIN WAVE CAPTURE METHODS

EEG is typically recorded by contact electrodes with conductive paste, while MEG detectors are in an array slightly removed from the head.  Remote detection of brain rhythms by electrical impedance sensors is described. [170]  Though non-contact is the only remote descriptor for EEG, this same detector design is applied to monitoring electrocardiogram with wrist sensor location. [171]  Passive brain wave fields extend as far as 12 feet from man as detected by a cryogenic antenna. [172]  This device is entirely adaptable to clandestine applications, and pointed comments are made on the disappearance of physiological remote sensing literature since the 1970’s for animals and humans, while all other categories of remote sensing research greatly expanded. [173] 

In 1976, the Malech Patent # 3951134  “Apparatus and method for remotely monitoring and altering brain waves” was granted.[174]  Example of operation is at 100 and 210 MHz, which are frequencies penetrating obstruction. [175]  “The individual components of the system for monitoring and controlling brain wave activity may be of conventional type commonly employed in radar”; and  “The system permits medical diagnosis of patients, inaccessible to physicians, from remote stations” are quotes indicating remote capacity.  License is to Dorne & Margolin Inc., but now protection is expired with public domain.  The Malech patent utilizes interference of 210 and 100 MHz frequencies resulting in a 110 MHz return signal, from which EEG waveform is demodulated. 

 A capability for ‘remote EEG’ is predicted by electromagnetic scattering theory using ultrashort pulses, [176] which is different from the unpulsed Malech patent.  Sampling rate for EEG specific concept recognition is only 1000 Hz (103/sec.), 6 compared to common radio frequency technology available at picosecond pulse widths allowing a considerably higher sampling rate.  Current review of microscopic electric field imaging describes phase, amplitude, and polarization changes of reflected waveform that would be expected to propagate at distance, and be detectable.[177]

            In addition to the radio frequency ultrashort pulse and interference methods above that are compatible with target tracking radars, the capacity to detect remote electric field changes is evident in present Radio Frequency Identification Device (RFID) technology.   Passive and semi-passive RFID tags encode information by electrically induced impedance changes that modulate the power of the backscattered ‘echo’ according to the equation: [178]

PS = I2 . Rr

 

                                    Where: PS = Power reflected by an antenna.

                                                 I   = Current.

                                                 Rr = Radiation resistance of antenna (without current).

 

Semi-passive RFID tags have a battery supplying current that modulates the backscatter mechanism with present read ranges as far as 30.5 meters under commercial reader power regulations, [179] but military capabilities considered include targeting of RFID tags by missiles. [180]  Though the above equation stipulates antenna properties, the human body is regarded as an antenna by several treatments. [181] [182] [183] [184]  Comparing the occurrence of human EEG current as on the order of microamps [185] to descriptions of RFID practical operation from 2.5-25 microamps, 178 and at the nanoamp level [186] provides support for design approaches to gaining radar encephalographic information by this mechanism.  The electrically modulated scatterer literature dates back to 1955. [187] 

            A dissertation exists on microwave detection of neural activity in cockroaches [188] along with related work apparently presented at a symposium, [189] and a portable diagnostic microwave patent for a detector of neural activity that describes animal studies. [190]  Though the dissertation suggests electro-mechanical impedance changes, the patent indicates that there is more direct electrical modulation of microwave backscatter, and operation at continuous wave.  Though not developed for remote application, this system has some correspondence to RFID electric field modulated backscatter methods. Review of the feasibility for imaging brain neuronal electrical activity by methods potentially capable of rather stringent resolution requirements considers other approaches to bioelectric field imaging. [191]

 

THOUGHT READING COVERT DEVELOPMENT EVIDENCE

The research arm of agencies mandated to covertly acquire information would certainly develop to operational capability any thought reading potential, which was reported feasible 30 years ago to the Department of Advanced Research Projects Agency (DARPA).  Reports that such development has progressed are multiple, and two are confirmed by details of the 1975 DARPA EEG specific word recognition report, which itself is evidence of development covert to open databases. 2  An International Committee of the Red Cross Symposium synopsis states EEG computer mind reading development by Lawrence Pinneo in 1974 at Stanford. [192]  A letter by the Department of Defense Assistant General Counsel for Manpower, Health, and Public Affairs, Robert L. Gilliat affirmed brain wave reading by the Advanced Research Projects Agency in 1976, [193] the same year as the Malech remote EEG patent grant.  Neglect of developing such a capacity by security institutions in the 22 years between the Pinneo report and relatively recent confirmations is not credible.  The further Dickhaut, 1998 Government report 3 appears more advanced than the journal literature, while the National Technical Information Service’s database is only clumsily searchable with availability limited by charge of commercial copyright rates for public information.

Dr. John Norseen of Lockheed Martin Aeronautics is quoted in news articles that thought reading is possible and has had development.[194] [195]  At least knowledge of Dickhaut, 1998 3 is evidenced in reference by a Norseen presentation, [196] but he predicted by 2005 the deployment of thought reading detectors for profiling terrorists at airports. 195  A further acknowledgement of developing a device to read terrorists’ minds at airports was made in a NASA presentation to Northwest Airlines security specialists. [197]  Statements in all news articles indicate remoteness of brain wave detection, though somewhat proximate. 

“Thought reading or synthetic telepathy” communications technology procurement is considered in a 1993 Jane’s[h] Special Operations Forces (SOF) article:  “One day, SOF commandos may be capable of communicating through thought processes.” [198]  Descriptive terms are “mental weaponry and psychic warfare.”  Although contemplated in future context, implied is availability of a technology with limited mobility, since troop deployment anticipation must assume prior development.  Victim complaints that mind reading is part of an assault upon them are very similar to such a capacity.  Other complaints by these victims, such as technologically transmitted voice assault are upheld by considerable documentation that individually isolated voice transmission is feasible, even at a distance and within structures, 175 and a presumptive diagnosis of such complaints is largely consistent with microwave exposure [199]--a basis for both internal voice and EEG capture technologies. 

 

DISCUSSION

There is considerable confirmation of an ability to recognize specific concepts by brain activity across subjects at some level of accuracy.  Identifying visual images viewed by a subject solely by measures of mental activity is replicated across seven groups by either EEG or fMRI.  Five groups report success in visually viewed word identification by brain activity in these methods.  Isolated groups report EEG word recognition by auditory perception, prior to vocalization, or as independently recalled.  Although many studies examine lesser sets of concepts, when added together the collective differentiation of these smaller sets approaches 100, and recent reports 16 21 differentiate even larger comparison numbers with a report of effective visual cortex image decoding. 22  In all, ten separate groups report some level of specific concept recognition by EEG, MEG, or fMRI.  Word category distinctions are expected from these specific differences.  EEG, MEG, PET, or fMRI techniques discriminate some 42 word class or dimension distinctions, many of which would survive separate direct comparison just by reported results.   

Considerable capacity to specifically detect and differentiate other mental states is evident from literature reports by EEG.  The fact that EEG signals are detected on a voluntary unprompted basis for turning on computer programs, 166 playing Pac Man, 162 and robot guidance 163 164 165 suggests the feasibility of a similar capacity for specific EEG concept recognition.  Although most concept recognition work is related to stimulus prompted responses, unprompted detection of numbers apparently as a class, has limited report. 168 

The finding that words can be classified by superposition of sine waves 8 or by frequency sub-waveform topography 3 suggests an obvious interpretation, when considering word category blood flow activations of cell assemblies. 195  The frequencies resulting from neuron firing rates in the distributed, yet somewhat discrete regions, when interference phase summed and subtracted by arrival from different locations results in word representation in the brain’s language.  Such results and the fact that the best recognition rates for words are obtained by the difference between an electrode pair 6 3supports the concept that a resultant waveform would provide similar information.         

Remote electric field determination of such a resultant waveform for decoding the encephalogram does have covert development indication by news reports.  The potential for thought reading and such a remote capacity is cautioned by French government scientific panel. [200]  At various levels of remoteness numerous methods or potentially exploitable mechanisms for detecting brainwave activity are described in open literature. 

Complete rejection of reports of a remote mind reading capability is just as presumptuous, in the face of complaints, as has been the dismissal of remote voice transmission capacity. 175  News reports of covert thought reading development have some confirmation in the Pinneo 1975 study 2 and Dickhaut, 1998 3 with independent news assertions of somewhat remote thought reading development “against terrorists” affirming each other.  Special operations officials consider procurement of a similar remote capacity to that of which many victims complain.  Though victims will regard their experience to affirm such a thought reading capability, professional prejudice classifies such complaints as within Schneiderian symptoms defining psychiatric condition.  The certain fact is that these claims have had no adequate investigation, and the available evidence questions the routinely egregious denial of civil rights to such individuals.  Complaints involving mind reading must at least receive rational investigation rather than ignorant professional dismissal convenient to practice with lucrative livelihood benefit. 

It is known that government elements have done work in thought reading development.  The logic that in the 30 years since the Pinneo work started, this capacity is operationally applied is too sound to dismiss victim corroboration and other evidence.  Funding for projects by the defense and security agencies is considerably greater than for open science, and thought reading would unquestionably be a priority area.  Except for the evidence for misuse as conjuncted with another radio frequency communications technology, [201] 199 [202]  and numerous obvious indications of such a capacity freely available, this author would prefer the information remain classified.  Particularly disturbing is the existence of remote electric field determination methods in the public domain.  Educated democracies should not be complacent at any prospect of mind reading, given the potential for privacy loss, civil rights violation, and political control.

 

Acknowledgements:  Thanks are given to God for inspiration and guidance as well as John Allman, Secretary of Christians Against Mental Slavery for invaluable materials and support (website http://www.slavery.org.uk/ ).

 

REFERENCES


[a] This article has been partly supported by substantial financial contributions from Christians Against Mental Slavery http://www.slavery.org.uk 

[b] Since the copyright date this article is updated in the Specific Concept Recognition, as well as the Proximate and Remote Brain Wave Capture Methods sections.

[c] Pinneo’s report does not include all experiments reported to the Defense Advanced Research Projects Agency in the six annual reports over the 3 year contract.

[d] Over the experiments presented by the report, chance would be from 6.5 to 14% depending on the size of tested vocabulary.

[e] Suppes points out that this may have been due to increased averaging per se.

[f] Though apparently only single electrodes or pairs were utilized for prediction, the best recognition rates were not always from the same electrode or pair.

[g] Almost half of the Pinneo report is devoted to resolving such confounds.

[h] Jane’s is the most respected and authoritative of defense reporting services.



[1] The Bible Job 42: 2, Psalms 139: 2,  94: 11, I Chronicles 28: 9, Isaiah 66: 18.

[2] Pinneo LR and Hall DJ. “Feasibility Study for Design of a Biocybernetic Communication System” Report #ADA017405 National Technical Information Service (NTIS), 1975.  Prepared for the Advanced Research Projects Agency Order #2034, Program Code #2D20, Contractor: Stanford Research Institute Contract dates: 2/9/72-8/31/76, SRI Project LSU-1936. (US cost ~$50.)  vailable at http://www.slavery.org.uk/Pinneo.doc 

[3] Dickhaut RH. Neuroelectric activity and analysis in support of direct brainwave to computer interface development. National Technical Information Service report # ADA350432, 1998. Available at: http://www.slavery.org.uk/spectra.pdf & http://www.freedomfchs.com/mindreadingspectra.pdf

[4] Suppes P, Lu Z, and Han B.  “Brain wave recognition of words” Proc Natl Acad Sci 94: 14965-69, 1997.  Printable free online thru Pubmed or at  http://www.pnas.org/cgi/content/full/94/26/14965

[5] Suppes P, Han B, and Lu Z. “Brain-wave recognition of sentences” Proc Natl Acad Sci 95: 15861-66, 1998.  Printable free online thru Pubmed or at  http://www.pnas.org/cgi/content/full/95/26/15861

[6] Suppes P, Han B, Epelboim J, and Lu Z. “Invariance of brain-wave representations of simple visual images and their names” Proc Natl  Acad  Sci 96: 14658-63, 1999.  Printable free online thru Pubmed or at http://www.pnas.org/cgi/content/full/96/25/14658

[7] Suppes P, Han B, Epelboim J, and Lu ZL. “Invariance between subjects of brain wave representations of language” Proc Natl Acad  Sci 96(22): 12953-8, 1999.  Printable free online thru PubMed or at http://www.pnas.org/cgi/content/full/96/22/12953

[8] Suppes P and Han B. “Brain-wave representation of words by superposition of a few sine waves” Proc Natl Acad Sci 97: 8738-43, 2000.  Printable free online thru Pubmed or at http://www.pnas.org/cgi/content/full/97/15/8738

[9] de Barros JA, Carvalhaes CG, de Mendonca JPRF, and Suppes P. “Recognition of Words from the EEG Laplacian” Brazilian Journal of Biomedical Engineering 21(2-3): 143-50, 2005. At http://www.ecologia.ufjf.br/admin/upload/File/Barros%20et%20al.,%202006.pdf

[10] Kim M-J, Shin S-C, Song Y, and Ryu CS. “Yes/No Discrimination With Spatio-Temporal Characteristics of EEG” Proceedings of the 23rd International Conference of the IEEE Engineering in Medicine and Biology Society 2: 2009-12, 2001.  Abstract at http://ieeexplore.ieee.org/xpl/freeabs_all.jsp?arnumber=1020625 

[11] Assadullahi R and Pulvermuller F. “Neural Network Classification of Word Evoked Neuromagnetic Brain Activity” In: Wermter S, Austin J, and  Willahaw D (eds.) Lecture Notes in Artificial Intelligence: Emergent Neurocomputational Architechures Based on Neuroscience Heidelberg Springer, p 311-20, 2001.  More limited preliminary communication at http://www.his.sunderland.ac.uk/durhamab/ramin.doc

[12] Guimares MP, Wong DK, Uy ET, Grosenick L, and Suppes P. “Single-Trial Classification of MEG Recordings” IEEE Trans Biomedical Engineering 54(3): 436-43, 2007.

[13]  Selden G. “Machines That Read Minds” Sci Digest Oct 89: 60-6, 1981.   

[14] Kiyuna T, Tanigawa T, and Yamazaki T. Patent #5785653 “System and method for predicting internal condition of live body” USPTO granted 7/28/98.

[15] Mardirossian A. Patent #6011991 “Communication system and method including brain wave analysis and/or use of brain activity” USPTO granted 1/4/00.

[16] Mitchell TM, Shinkareva SV, Carlson A, Chang KM, Malave VL, Mason RA, and Just MA. “Predicting Human Brain Activity Associated with the Meaning of Nouns” Science 320: 1191-5, 2008.

[17] Haxby JV, Gobbini MI, Furey ML, Ishai A, Schonten JL, and Pietrini P. “Distributed and Overlapping Representations of Faces and Objects in Ventral Temporal Cortex” Science 293(5529): 2425-30, 2001.

[18] Spiridon M, and Kanwisher N. “How Distributed is Visual Category Information in Human Occipito-Temporal Cortex? An fMRI Study” Neuron 35: 1157-65, 2002.

[19] Cox DD, and Savoy RL. “Functional magnetic resonance imaging (fMRI) “brain reading”: detecting and classifying distributed patterns of fMRI activity in human visual cortex” Neuroimage 19: 261-70, 2003.

[20] Shinkareva SV, Mason RA, Malave VL, Wang W, Mitchell TM, and Just MA. “Using fMRI Brain Activation to Identify Cognitive States Associated with Perception of Tools and Dwellings” PLoS ONE. 2008; 3(1): e1394.  Published online 2008 January 2. doi: 10.1371/journal.pone.0001394 at http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=2148074

[21] Kay KN, Naselaris T, Pernger RJ, and Gallant JL. “Identifying natural images from human brain activity” Nature 452: 352-6, 2008.

[22] Miyawaki Y, Uchida H, Yamashita O, Sato M, Morito Y, Tanabe HC, Sadato N, Kamitani Y. Visual image reconstruction from human brain activity using a combination of multiscale local image decoders. Neuron 2008;60:915-29

[23] Haynes JD and Rees G. “Decoding mental states from brain activity in humans” Nature Neuroscience Rev 7: 523-34, 2006.

[24] Norman KA, Polyn SM, Detre GJ, and Haxby JV. “Beyond mind-reading: multi-voxel pattern analysis of fMRI data” Trends Cognit Sci 10(9): 423-30, 2006.

[25] Thirion B, Duchesnay E, Hubbard E, Dubois J, Poline JB, Lebihan D, and Dehaene S. “Inverse retinotopy: Inferring the visual content of images from brain activation patterns” Neuroimage 33: 1104-16, 2006.

[26] Vandenberghe R, Price C, Wise R, Josephs O, and Frackowiak RSJ. “Functional anatomy of a common semantic system for words and pictures” Nature 383: 354-6, 1996.

[27] Chao LL, Haxby JV, and Martin A. “Attribute-based neural substrates in temporal cortex for perceiving and knowing about objects” Nature Neurosci  2(10): 913-9, 1999.

[28] Moore CJ and Price CJ. “Three Distinct Ventral Occipitotemporal Regions for Reading and Object Naming” NeuroImage 10: 181-92, 1999.

[29] Le Clec’H G, Dehaene S, Cohen L, Mehler E, Dupoux E, Poline JB, Lehericy S, van de Moortele PF, and Le Bihan D. “Distinct Cortical Areas for Names of Numbers and Body Parts Independent of Language and Input modality” NeuroImage 12: 381-91, 2000.

[30] Chee MWL, O’Craven KM, Bergida R, Rosen BR, and Savoy RL. “Auditory and Visual Word Processing Studied With fMRI” Hum Brain Mapp 7: 15-28, 1999.

[31] Isselbacher KJ, Adams RD, Brunwald E, Petersdorf RG, and Wilson JD (eds.) Harrison’s Principles of Internal Medicine 9th edition, McGraw-Hill, p 141-2, 1980.

[32] Warrington EK and Shallice T. “Category specific semantic impairments” Brain 107(Pt 3): 829-54, 1984.

[33] Damasio H, Grabowski TJ, Tranel D, Hichwa RD, and Damasio AR. “A neural basis for lexical retrieval” Nature 380: 499-505, 1996.

[34] De Renzi E. “Disorders of Visual Recognition” Semin Neurol 20(4): 479-85, 2000.

[35] Pulvermuller F. “Words in the brain’s language” Behav Brain Sci 22: 253-336, 1999.

[36] Khader P and Rosler F. “EEG power and coherence analysis of visually presented nouns and verbs reveals left frontal processing differences” Neurosci Lett 354: 111-14, 2004.

[37] Preissl H, Pulvermuller F, Lutzenberger W, and Birbaumer N. “Evoked potentials distinguish between nouns and verbs” Neurosci Lett 197: 81-3, 1995.

[38] Kellenbach ML, Wijers AA, Hovius M, Mulder J, and Mulder G. “Neural Differentiation of Lexico-Syntactic Categories or Semantic Features? Event-Related Potential Evidence for Both” J Cog Neurosci 14(4): 561-77, 2002.

[39] Khader P, Scherag A, Streb J, and Rosler F. “Differences between noun and verb processing in minimal phrase context: a semantic priming study using event-related brain potentials” Cogn Brain Res 17: 293-313, 2003.

[40] Federmeier KD, Segal JB, Lombrozo T, and Kutas M. “Brain responses to nouns, verbs and class-ambiguous words in context” Brain 123(12): 2552-66, 2000. 

[41] Brown WS, Lehmann D, and Marsh JT. “Linguistic Meaning Related Differences in Evoked Potential Topography: English, Swiss-German, and Imagined” Brain Lang 11: 340-53, 1980.

[42] Pulvermuller F, Mohr B, and Schleichert H. “Semantic or lexico-syntactic factors: what determines word-class specific activity in the human brain?” Neurosci Lett 275: 81-4, 1999.

[43] Pulvermuller F, Lutzenberger W, and Preissl H. “Nouns and Verbs in the Intact Brain: Evidence from Event-retlated Potentials and High-frequency Cortical Responses” Cerebral Cortex 9(5): 497-506, 1999.  

[44] Pulvermuller F, Preissl H, Lutzenberger W, and Birbaumer N. “Brain Rhythms of Language: Nouns Versus Verbs” Eur J Neurosci 8: 917-41, 1996.

[45] Pulvermuller F, Harle M, and Hummel F. “Walking or Talking? Behavioral and Neruophysiological Correlates of Action Verb Processing” Brain Lang 78: 143-68, 2001.

[46] Pulvermuller F, Harle M, and Hummel F. “Neurophysiological distinction of verb categories” Cog Neurosci 11(12): 2789-93, 2000.

[47] Lavric A, Pizzagalli D, Forstmeier S, and Rippon G. “A double dissociation of English past-tense production revealed by event-related potentials and low-resolution electromagnetic tomography (LORETA)” Clin Neurophysiol 112: 1833-49, 2001.

[48] Dhond RP, Marinkovic K, Dale AM, Wotzel T, and Halgren E. “Spatiotemporal maps of past-tense verb inflection” Neuroimage 19: 91-100, 2003.

[49] Weyerts H, Munte TF, Smid HGOM, and Heinze H-J. “Mental representations of morphologically complex words: and event-related potential study with adult humans” Neurosci Lett 206: 125-8, 1996.

[50] Munte TF, Say T, Clahsen H, Schlitz K, and Kutas M. “Decomposition of morphologically complex words in English: evidence from event-related potentials” Cogn Brain Res 7: 241-53, 1999.

[51] Weyerts H, Penke M, Dohrn U, Clahsen H, and Munte TF. “Brain potentials indicate differences between regular and irregular German plurals” Neuroreport  8(4): 957-62, 1997.

[52] Penke M, Weyerts H, Gross M, Zander E, Munte TF, and Clahsen H. “How the brain processes complex words: an event-related potential study of German verb inflection” Cogn Brain Res 6: 37-52, 1997.

[53] Gross M, Say T, Kleingers M, Clahsen H, and Munte TF. “Human brain potentials to violations in morphologically complex Italian words” Neurosci Lett 241: 83-6, 1998.

[54] Kounios J and Holcomb PJ. “Concreteness Effects in Semantic Processing: ERP Evidence Supporting Dual-Coding Theory” J Exp Psychol 20(4): 804-23, 1994.

[55] West CW and Holcomb PJ. “Imaginal, Semantic, and Surface-Level Processing of Concrete and Abstract Words: An Electrophysiological Investigation” J Cogn Neurosci  12: 1024-37, 2000.

[56] Weiss S and Rappelsberger P. “EEG coherence within the 13-18 band as a correlate of a distinct lexical organization of concrete and abstract nouns in humans” Neurosci Lett 209: 17-20, 1996.

[57] Schack B, Weiss S, and Rappelsberger P. “Cerebral Information Transfer During Word Processing: Where and When Does It Occur and How Fast is it?” Hum Brain Mapp 19: 18-36, 2003.

[58] Weiss S and Rappelsberger P. “Left Frontal EEG Coherence Reflects Modality Independent Language Processes” Brain Topogr 11(1): 33-42, 1998.

[59] Weiss S, Muller HM, and Rappelsberger P. “Theta synchronization predicts efficient memory encoding of concrete and abstract nouns” NeuroReport 11(11): 2357-61, 2000.

[60] Neville HJ, Mills D, and Lawson DS. “Fractionating Language: Different Neural Subsystems with Different Sensitive Periods” Cerebral Cortex 2: 244-58, 1992.

[61] Munte TF, Wieringa BM, Weyerts H, Szentkuti A, Matzke M, and Johannes S. “Differences in brain potentials to open and closed class words: class and frequency effects” Neuropsychologia 39: 91-102, 2001.

[62] Van Petten C and Kutas M. “Influences of semantic and syntactic contex on open- and closed-class words” Mem Cogn 19: 95-112, 1991.

[63] Pulvermuller F, Lutzenberger W, and Birbaumer N. “Electrocortical distinction of vocabulary types” Electroenceph Clin Neurophysiol 94: 357-70, 1995.

[64] Mohr B, Pulvermuller F, and Zaidel E. “Lexical Decision After Left, Right, and Bilateral Presentation of Function Words, Content Words, and Non-Words: Evidence For Interhemispheric Interaction” Neuropsychologia 32(1): 105-24, 1994.

[65] Pulvermuller F, Assedollahi R, and Ekbert T. “Neuromagnetic evidence for early semantic access in word recognition” J Neurosci 13: 201-5, 2001.

[66] Muller HM and Kutas M. “What’s in a name? Electrophysiological differences between spoken nouns, proper names and one’s own name” Neuroreport 8: 221-5, 1996.

[67] Dehaene S. “Electrophysiological evidence for category-specific word processing” Neuroreport 6: 2153-7, 1995.

[68] Antal A, Keri S, Kovacs G, Janka Z, and Benedek G. “Early and Late components of visual categorization:  an evant-related potential study” Cogn Brain Res 9: 117-19, 2000.

[69] Antal A, Keri S, Kovacs G, Liszli P, Janka Z, and Benedek G. “Event-related potentials from a visual categorization task” Brain Res Protocols 7: 131-6, 2001.

[70] Ji J, Porjesz B, and Begleiter H. “ERP components in category matching tasks” Electroencephalogr Clin Neurophysiol 108: 380-9, 1998.

[71] Thorpe S, Fize D, and Marlot C. “Speed of processing in the human visual system” Nature 381: 520-2, 1996.

[72] Kiefer M. “Perceptual and semantic sources of category-specific effects: Event-related potentials during picture and word categorization” Mem Cogn 29(1): 100-16, 2001.

[73] Skrandies W and Chiu MJ. “Dimensions of affective meaning – behavioral evoked potential correlates in Chinese subjects” Neurosci Lett 341: 45-8, 2003.

[74] Skrandies W. “Evoked potential correlates of semantic meaning—A brain mapping study” Cog Brain Res 6: 175-83, 1998.

[75] Schapkin SA, Gusev AN, and Kuhl J. “Categorization of unilaterally presented emotional words: an ERP analysis” Acta Neurobiol Exp 60: 17-28, 2000.

[76] Dietrich DE, Waller C, Johannes S, Wieringa BM, Emrich HM, and Munte TF. “Differential Effects of Emotional Content on Event-Related Potentials in Word Recognition Memory” Neuropsychobiol 43: 96-101, 2001.

[77] Krause CM, Korpilahti P, Porn B, Joskim J, and Lang HA. “Automatic auditory word perception as measured by 40 Hz EEG responses” Electroencephal Clin Neurophysiol 107: 84-7, 1998.

[78] Diesch E, Biermann S, and Luce T. “The magnetic mismatch field elicited by words and phonological non-words” Neuroreport 9(3): 455-60, 1998.

[79] Lutzenberger W, Pulvermuller F, and Birbaumer N. “Words and pseudowords elicit distinct patterns of 30-Hz EEG responses” Neurosci Lett 176: 115-18, 1994.

[80] Kiefer M. “Perceptual and semantic sources of category-specific effects: Event-related potentials during picture and word categorization” Mem Cog 29(1): 100-16, 2001.

[81] Assadollahi R and Pulvermuller F. “Neuromagnetic evidence for early access to cognitive representations” Cog Neurosci Neurophysiol 12(2): 207-13, 2001.

[82] Steinhauer K. “Electrophysiological correlates of prosody and punctuation” Brain Lang 86: 142-64, 2003.

[83] Lange JJ, Wijers AA, Mulder LJM, and Mulder G. “Color selection and location selection in ERPs: differences, similarities and ‘neural specificity’” Biol Psychology 48: 53-82, 1998.

[84] Mouchetant-Rostaing Y, Girad M-H, Benlin S, Aguera P-E, and Pernier J. “Neurophysiological correlatres of face gender processing in humans” Eur J Neurosci 12: 303-12, 2000.

[85] Aguirre GK, Zarahn E, and D’Esposito M. “An Area within the Human Ventral Cortex Sensitive to “Building” Stimuli: Evidence and Implications” Neuron 21: 373-83, 1998.

[86] Epstein R, Harris A, Stanley D, and Kanwisher N.  “The Parahippocampal Place Area: Recognition, Navigation, or Encoding?” Neuron 23: 115-25, 1999.

[87] Kanwisher N, McDermott J, and Chun MM. “The Fusiform Face Area: A Module in Human Extrastriate Cortex Specialized for Face Perception” J Neurosci  17(11): 4302-11, 1997.

[88] Gauthier I, Scudlarski P, Gore JC, and Anderson AW. “Expertise for cars and birds recruits brain areas involved in face recognition” Nature Neurosci 3(2): 191-7, 2000.

[89] Ishai A, Ungerleider LG, Martin A, Schouten JL, and Haxby JV. “Distributed representations of objects in the human ventral visual pathway” Proc Natl Acad Sci 96(16): 9379-84, 1999. 

[90] Martin A. “Functional Neuroimaging of Semantic Memory” In: Cabeza R and Kingstone A (eds.) Handbook of Functional Neuroimaging of Cognition MIT Press, Cambridge, Mass p 153-86, 2001.

[91] Martin A and Chao LL. “Semantic memory and the brain: structure and processes” Curr Opinion Neurobiol 11(2): 194-201, 2001.

[92] Small SL, Hart J, Nguyen T, and Gordon H. “Distributed representations of semantic knowledge in the brain” Brain 118: 441-53, 1995.

[93] Joseph JE. “Functional Neuroimaging Studies of Category Specificity in Object Recognition: A critical review and meta-analysis” Cog Affect Behav Neurosci 1(2): 119-36, 2001.

[94] Warburton E, Wise RJS, Price CJ, Weiller C, Hadar U, Ramsay S, and Frackowiak RSJ. “Noun and verb retrieval by normal subjects. Studies with PET” Brain 119(Pt 1): 159-79, 1996.

[95] Perani D, Cappa SF, Schnur T, Tettamanti M, Collina S, Rosa MM, and Faziol F. “The neural correlates of verb and noun processing: a PET study” Brain 122(12): 2237-44, 1999.  Also at http://brain.oupjournals.org/cgi/content/full/122/12/2337

[96] Tyler LK, Russell R, Fadili J, and Moss HE. “The neural representation of nouns and verbs: PET studies” Brain 124(8): 1619-34, 2001.

[97] Beretta A, Campbell C, Carr TH, Huang J, Schmitt LM, Cristianson K, and Cao Y.  “An ER-fMRI investigation of morphological inflection in German reveals that the brain makes a distinction between regular and irregular forms” Brain Lang 85: 67-92, 2003.

[98] Jaeger JJ, Lockwood AH, Kemmerer DL, Val Valin RD, Murphy BW, and  Khalak HG. “A Positron Emission Tomographic Study of Regular and Irregular Verb Morphology in English” Language 42(3): 451-97, 1996.

[99] Damasio H, Grabowski TJ, Tranel D, Ponto LLB, Hichwa RD, and Damasio AR. “Neural Correlates of Naming Actions and of Naming Spatial Relations” NeuroImage 13: 1053- 64, 2001.

[100] Moscovitch M, Kapur S, Kohler S, and Houle S. “Distnict neural correlates of visual long-term memory for spatial location and object identity: a positron emission tomography study in humans” Proc Natl Acad Sci 92: 3721-5, 1995.  Also at http://www.pnas.org/cgi/reprint/92/9/3721.pdf

[101] Emmorey K, Damasio H, McCullough S, Grabowski T, Ponto LLB, Hichwa RD, and Bellugi U. “Neural Systems Underlying Spatial Language in American Sign Language” Neuroimage 17: 812-24, 2002.

[102] Mummery CJ, Patterson K, Hodges JR, and Price CJ. “Functional Neuroanatomy of the Semantic System: Divisible by What?” J Cogn Neurosci 10(6): 766-77, 1998.

[103] Martin A, Haxby JV, Lalonde FM, Wiggs CL, and Ungerleider LG. “Discrete Cortical Regions Associated with Knowledge of Color and Knowledge of Action” Science 270: 102-5, 1995.

[104] Chao LL and Martin A. “Cortical Regions Associated with Perceiving, Naming, and Knowing about Colors” J Cogn Neurosci 11(1): 25-35, 1999.

[105] Kounios J, Koenig P, Glosser G, DeVita C, Dennis K, Moore P, and Grossman M. “Category-specific medial temporal lobe activation and the consolidation of semantic memory: evidence from fMRI” Cogn Brain Res 17: 484-94, 2003.

[106] Beauregard M, Chertkow H, Bub D, Murtha S, Dixon R, and Evans A. “The Neural Substrate for Concrete, Abstract, and Emotional Word Lexica: A Positron Emission Tomographic Study” J Cogn Neurosci 9(4): 441-61, 1997.

[107] Fiebach CJ and Friederici AD. “Processing concrete words: fMRI evidence against a specific right-hemisphere involvement” Neuropsychologia 42: 62-70, 2003.

[108] Kiehl KA, Liddle PF, Smith AM, Mendrek A, Forster BB, and Hare RD. “Neural Pathways Involved in the Processing of Concrete and Abstract Words” Hum Brain Mapp 7: 225-33, 1999.

[109] Mellet E, Tzourio N, Denis M, and Mazoyer B. “Cortical Anatomy of mental imagery of concrete nouns based on their dictionary definition”  NeuroReport 9: 803-8, 1998.

[110] Grossman M, Koenig P, DeVita C, Glosser G, Alsop D, Detre J, and Gee J. “The Neural Basis for Category Specific Knowledge: An fMRI Study” NeuroImage 15: 936-48, 2002.

[111] Jessen F, Heun R, Erb M, Granath D-O, Klose U, Papassotiropoulos A, and Grodd W. “The Concreteness Effect: Evidence for Dual Coding and Context Availability” Brain Lang 74: 103-12, 2000.

[112] Grossman M, Koenig P, DeVita C, Glosser G, Alsop D, Debre J, and Gee J. “Neural Representation of Verb Meaning: An fMRI Study” Hum Brain Mapp 15: 124-34, 2002.

[113] Damasio H, Grabowski TJ, Tranel D, Hichwa RD, and Damasio AR. “A neural basis for lexical retrieval” Nature 380: 499-505, 1996.

[114] Grabowski TJ, Damasio H, and Damasio AR. “Premotor and Prefrontal Correlates of Category-Related Lexical Retrieval” NeuroImage 7: 232-43, 1998.

[115] Gorno-Tempini ML, Price CJ, Josephs R, Vandenberghe R, Cappa SF, Kapur N, and Frackowiak RSJ. “The neural systems sustaining face and proper name processing” Brain 121: 2103-18, 1998.  Also at http://brain.oupjournals.org/cgi/reprint/121/11/2103.pdf

[116] Gorno-Tempini ML and Price CJ. “Identification of famous faces and buildings: A functional neuroimaging study of semantically unique items” Brain 124: 2087-97, 2001.  Also at http://brain.oupjournals.org/cgi/content/full/124/10/2087

[117] Ishai A, Ungerleider LG, Martin A, and Haxby JV. “The Representation of Objects in the Human Occipital and Temporal Cortex” J Cogn Neurosci 12(Supp 3): 35-52, 2000.

[118] Ishai A, Ungerleider LG, and Haxby JV. “Distributed Neural Systems for the Generation of Visual Images” Neuron 28: 979-90, 2000.

[119] Ishai A, Haxby JV, and Ungerleider LG. “Visual Imagery of Famous Faces: Effects of Memory and Attention Revealed by fMRI” NeuroImage 17: 1729-41, 2002.

[120] Spitzer M, Kwong KK, Kennedy W, Rosen BR, and Belliveau JW. “Category-specific brain activation in fMRI during picture naming” NeuroReport  6: 2109-12, 1995.

[121] Mitchell TM, Hutchinson R, Just MA, Niculescu RS, Percira F, and Wang X. “Classifying Instantaneous Cognitive States from fMRI Data” Am Med Informatics Assoc, November, 2003.  Also at http://www-2.cs.cmu.edu/~tom/amia2003-final.pdf

[122] Joseph JE, Gathers AD, and Piper GA. “Shared and dissociated cortical regions for object and letter processing” Cogn Brain Res 17: 56-67, 2003.

[123] Polk TA and Farah MJ. “The neural development and orgainization of letter recognition: Evidence from functional neuroimaging, computational modeling, and behavioral studies” Proc Natl Acad Sci 95(3): 847-52, 1998.  Also at http://www.pnas.org/cgi/content/full/95/3/847

[124] Polk TA, Stallcup M, Aguirre GK, Alsop DC, D’Esposito M, Detre JA, and Farah MJ. “Neural Specialization for Letter Recognition” J Cogn Neurosci  14(2): 145-59, 2002.

[125] Dehaene S, Spelke E, Pinel P, Stanescu R, and Tsivkin S. “Sources of Mathematical Thinking: Behavioral and Brain-Imaging Evidence” Science 284(5416): 970-4, 1999.

[126] Simon O, Mangin JF, Cohen L, Le Bihan D, and Dehaene S. “Topographical Layout of Hand, Eye, Calculation, and Language-Related Areas in the Human Parietal Lobe” Neuron 33: 475-87, 2002.

[127] Cochin F, Cohen L, van de Moortele PF, and Dehaene S. “Differential Contributions of the Left and Right Inferior Parietal Lobules to Number Processing” J Cogn Neurosci 11: 617-30, 1999.

[128] Davidson RJ. “Affective Style and Affective Disorders: Perspectives form Affective Neuroscience” Cogn Emot 13(3): 307-30, 1998.

[129] Davidson RJ. “Anterior electrophysiological asymmetries, emotion, and depression: Conceptural and methodological conundrums” Psychophysiol 35: 607-14, 1998.

[130] Papousek B and Schulter G. “Covariations of EEG asymmetries and emotional states indicate that activity at frontopolar locations is particularly affected by state factors” Psychophysiol 39: 350-60, 2002.

[131] Tomarkin AJ and Keener AD. “Frontal Brain Asymmetry and Depression: A Self-regulatory Perspective” Cogn Emotion 12(3): 387-420, 1998.

[132] Heller W and Nitschke JB. “The Puzzle of Regional Brain Activity in Depression and Anxiety: The Importance of Subtypes and Comorbidity:” Cogn Emotion 12(3): 421-47, 1998.

[133] Davidson RJ, Marshall JR, and Tomarkin AJ. “While a Phobic Waits: Regional Brain Electrical and Autonomic Activity in Social Phobics during Anticipation of Public Speaking” Biol Psychiatry 47: 85-95, 2000.

[134] Patton RE. Patent #6292688 “Method and apparatus for analyzing neurological response to emotion-inducing stimuli” USPTO granted 9/18/01.

[135] Pizzagalli D, Koenig T, Regard M, and Lehmann D. “Faces and emotions: brain electric field sources during covert emotional processing” Neuropsychologia 36(4): 323-32, 1998.

[136] Eger E, Jedynak A, Iwaki T, and Skrandies W. “Rapid extraction of emotional expression: evidence form evoked potential fields during brief presentation of face stimuli” Neuropsychologia 41: 808-17, 2001.

[137] Bocker KBE, Baas JMP, Kenemans JL, and Verbaten MN.  “Stimulus-preceding negativity induced by fear: a manifestation of affective anticipation” Int J Psychophysiol 43: 77-90, 2001.

[138] Kotani Y, Hiraku S, Suda K, and Aihara Y. “Effect of positive and negative emotion on stimulus-preceding negativity prior to feedback stimuli” Psychophysiol 38: 873-78, 2001.

[139] Kotani Y, Kishida S, Hiraku S, Suda K, Ishii M, and Aihara Y. “Effects of information and reward on stimulus-preceding negativity prior to feedback stimuli” Psychophysiol 40(5): 818-28, 2003.

[140] Baas JMP, Kenemans JL, Bocker JL, and Verbaten MN. “Threat-induced cortical processing and startle potentiation” Neuroreport 13(1): 133-7, 2002.

[141] Farwell LA and Donchin E. “The Truth Will Out: Interrogative Polygraphy (“Lie Detection”) With Event-Related Brain Potentials” Psychophysiology 28(5): 531-47, 1991.

[142] Johnson MM and Rosenfeld JP. “Oddball-evoked P300-based method of deception detection in the laboratory II. Utilization of non-selective activation of relevant knowledge” Int J Psychophysiol 12: 289-306, 1992.

[143] Rosenfeld JP, Ellwanger J, and Sweet J. “Detecting simulated amnesia with event-related brain potentials” Int J Psychophysiol 19: 1-11, 1995.

[144] Allen JJB and Iacono WG. “A Comparison of methods for the analysis of event-related potentials in deception detection” Psychophysiology 34: 234-40, 1997.

[145] Lorenz J, Kunze K, and Bromm B. “Differentiation of conversive sensory loss and malingering by P300 in a modified oddball task” Neuroreport 9: 187-91, 1998.

[146] Tardif HP, Barry RJ, and Johnstone SJ. “ Event-related potentials reveal processing differences in honest vs. malingered memory performance” Int J Psychophysiol 46: 147-58, 2002.

[147] Brain Fingerprinting Laboratories, Inc., 108 West Palm Drive, Fairfield, IA 52556 at http://www.brainwavescience.com/

[148] Farwell LA and Smith SS. “Using Brain MERMER Testing to Detect Knowledge Despite Efforts to Conceal” J Forensic Sci 46(1): 135-46, 2001.

[149] Farwell LA. “Two new twists on the truth detector: brain-wave detection of occupational information” Psychophysiology 29(4A): S3, 1992.

[150] Farwell LA. Patent #5363858 “Method and apparatus for multifaceted electroencephalographic response analysis” USPTO granted 11/15/94.

[151] Farwell LA and Conte FL. Patent #5406956 “Method and apparatus for truth detection” USPTO granted 4/18/95.

[152] Farwell LA and Conte FL. Patent #5467777 “Method for electroencephalographic information detection” USPTO granted 11/21/95.

[153] Crites SL, Cacioppo JT, Gardner WL, and Berntson GG. “Bioelectrical Echoes From Evaluative Categoriization: II. A Late Positive Brain Potential That Varies as a Function of Attitude Registration Rather Than Attitude Report” J Person Soc Psychol  68(6): 997-1013, 1995.

[154] Chen R and Hallett M.  “The Time Course of Changes in Motor Cortex Excitability Associated with Voluntary Movement” Can J Neurol Sci 26(3): 163-9, 1999.

[155] Deeke L. “Bereitschaftspotential as an indicator of movement preparation in supplementary motor area and motor cortex” Ciba Found Symp 182: 132-231, 1987.

[156] Cui RQ, Lang HW, and Deeke L. “Neuroimage of Voluntary Movement: Topography of the Bereitschaftspotential, a 64-Channel DC Current Source Density Study” NeuroImage 9: 124-34, 1999.

[157] Birch GE. “Initial On-Line Evaluations of the LF-ASD Brain-Computer Interface With Able Bodied and Spinal-Cord Subjects Using Imagined Voluntary Motor Potentials” IEEE Trans Neural Syst Rehabil Eng 10(4): 219-24, 2002.

[158] Penney WD, Roberts SJ, Curran EA, and Stokes MJ. “EEG-Based Communication: A Pattern Recognition Approach” IEEE Trans Rehab Eng (2): 214-15, 2000.

[159] Kostov A and Polak M. “Paralell Man-Machine Training in Development of EEG-Based Cursor Control” IEEE Trans Rehab Eng 8(2): 203-5, 2000.

[160] Guger C, Harkam W, Hertnacs C, and Pfurtscheller G. “Prosthetic Control by an EEG-based Brain-Computer Interface (BCI)” In: Bühler C and Knops H (eds.) Assistive Technology on the Threshold of the new Millennium  2003.  Also at http://www.gtec.at/research/Publications/aaate.pdf

[161] Tanaka K, Matsunaga K, and Wang HO. “Electroencephalogram-based Control of an Electric Wheelchair” IEEE Transactions on Robotics 21(4): 762-66, 2005.

[162] Krepki R, Blankertz B, Curio G, and Muller K-R. “The Berlin Brain-Computer Interface (BBCI) towards a new communication channel for online control of multimedia applications and computer games” 9th International Congress on Distributed Multimedia Systems, 2003.  At http://ida.first.fhg.de/publications/KreBlaCurMue03.pdf

[163] Millan JR. “Adaptive Brain Interfaces” Communications of the ACM 46(3): 74-80, 2003.  Abstract at http://www.idiap.ch/publications/millan-2003-comm-acm.bib.abs.html

[164] Millan JR and Mourifio J. “Asynchronous BCI and Local Neural Classifiers: An Overview of the Adaptive Brain Interface Project” IEEE Transactions on Neural Systems and Rehabilitation Engineering (Brain-Computer Interface Technology) 11(2): 159-61, 2003.  Article also at ftp://ftp.idiap.ch/pub/reports/2003/millan_2003_nsre.pdf

[165] Millan JR, Renkens F, Mourifio J, and Gerstner W. “Non-Invasive Brain-Actuated Control of a Mobile Robot” Proceedings of the 18th Joint International Conference on Artificial Intelligence Aug 9-15, in press, 2003.  Article also at ftp://ftp.idiap.ch/pub/reports/2003/millan_2003_ijcai.pdf

[166] Kaiser J, Perelmouter J, Iversen IH, Neumann N, Ghanayim N, Hinterberger T, Kubler A, Kotchonbey B, and Birbaumer N. “Self-initiation of EEG-based communication in paralysed patients” Clin Neurophysiol 112: 551-4, 2001.

[167] Carmens JM, Lebedev MA, Crist RE, O’Doherty JE, Santucci DM, Dimitrov DF, Patil PG, Henriques CS, and Nicolelis MAL. “Learning to Control a Brain-Machine Interface for Reaching and Grasping by Primates” Public Library of Science, Biology Oct 1(1) 2003. At http://www.plosbiology.org/archive/1545-7885/1/2/pdf/10.1371_journal.pbio.0000042-L.pdf 

[168] Roscher G, Pogrzeba G, Emde D, and Neubauer. “Application of a Multi-Processor System for Recognition of EEG-Activities in Amplitude, Time and Space in Real Time” In: D’Hollander EH, Joubert GR, Peters FJ, Trottenberg U (Eds.) Parallel Computing: Fundamentals, Applications and New Directions Elsevier Science B. V., p 89-96, 1998.  Abstract at http://www.icsroscher.de/Parco.htm

[169] Roscher G, Herrmann WM, Henning K, Wendt D, Fechner S, Godenschweger F, Weib C, Abel E, Rijhwani A, Martinez J, Karawas A, and Dahan N. “A System to Recognize, Estimate and Describe Single Events in the Spontaneous Electroencephalogram: Example for Single Sweep N1 and P2 Detection”  Clinical Applications of Advanced EEG Data Processing Rome, May 8-9 p 47, 1995.  At http://www.icsroscher.de/Rom.html

[170] Harland CJ, Clark TD, and Prance RJ. “Remote detection of human electroencephalograms using ultrahigh input impedance electric potential sensors” Appl Physics Lett  81(17): 3284-6, 2002. 

[171] Harland CJ, Clark TD, and Prance RJ. “High resolution ambulatory electrocardiographic monitoring using wrist mounted electric potential sensors” Meas Sci and Technol 14: 923-8, 2003.  Abstract at http://www.iop.org/EJ/abstract/0957-0233/14/7/305

[172] Taff BE and Stoller KP. Patent #4940058 “Cryogenic remote sensing physiograph” USPTO granted 7/10/90.

[173] Stoller KP and Taff BE. “Remote Physiological Sensing: Historical Perspective, Theories and Preliminary Developments” Med Instrum 20(5): 260-5, 1986.

[174] Malech RG. Patent #3951134 “Apparatus and method for remotely monitoring and altering brain waves” USPTO granted 4/20/76.

[175] McMurtrey JJ. “Inner Voice, Target Tracking, and Behavioral Influence Technologies” in press 2004.  Accessed 1/3/09 at http://www.slavery.org.uk/InnerVoiceTargTrackBehavInflu.doc 

[176] Department of the Army, USAF Scientific Advisory Board. “New World Vistas: air and space for the 21st century” 14 vol. (Ancillary Volume) p 89-90, 1996.  Also at http://mfhr.org/new_vistas.htm

[177] Rosner BT and van der Weide DW. “High-frequency Near-field Microscopy” Rev Sci Instruments 73(7): 2505-2525, 2002.

[178] Finkenzeller K (Translated by Waddington R). RFID Handbook [electronic resource].  Hoboken NJ; Wiley: p 123 & 139, 2003.

[179] Lahiei S. RFID Sourcebook Upper Sadle River, NJ: IBM Press; p  17, 2006.

[180] Juels A. “RFID Security and Privacy: A Research Survey” IEEE J Selected Areas in Communication 24(2): 381-94, 2006.

[181] Andersen JB and Hansen F. “Antennas for VHF/UHF Personal Radio: A Theoretical and Experimental Study of Characteristics and Performance” IEEE TransVehicular Technology 26(4): 349-57, 1977.

[182] Krupka Z. “The Effect of the Human Body on Radiation on Radiation Properties of Small-Sized Communication Systems” IEEE Trans Antennas and Propagation 16(2): 154-63, 1968.

[183] Andersen JB and Balling P. “Admittance and Radiation Efficiency of the Human Body in the Resonance Region” Proceedings of the IEEE 60(7): 900-1, 1972.

[184] Weinberger Z and Richter ED. “Cellular telephones and effects on the brain: The head as an antenna and brain tissue as a radio receiver” Med Hypotheses 59(6): 703-5, 2002.

[185] Romero G, Arenas MG, Castillo PA, and Merelo JJ. “Evolutionary Design of a Brain-Computer Interface” Lecture Notes in Computer Science, Berlin; Springer: vol 3512, p 669-76, 2005.

[186] Pillai V, Heinrich H, Dieska D, Nikitin PV, Martinez R, and Rao KVS. “An Ultra-Low Power Long Range Battery/Passive RFID Tag for UHF and Microwave Bands With a Current Consumption of 700 nA at 1.5 V” IEEE Trans Circuits And Systems 54(7): 1500-12, 2007. 

[187] Richmond JH. “Modulated Scattering Technique for the Measurement of Field Distributions” IRE Trans Microwave Theory & Technique 3: 13-15, 1955.

[188] Aaby TG. “Observation of Neural Activity as a Variation of Neural Activity (Cockroach)” Proquest Dissertations and Theses. Section 0128, Part 0176 118 pages; [Ph.D. dissertation] United States – Michigan State University; 1982. Publication Number: AAT 8308894.

[189] Rosenberg B and Aaby T. “Cardiac and Neural Activity Monitored by Microwave Scattering” :Presented at the 3rd Annual Symposium of the Bioelectromagnetics Society, Washington, DC,  Aug 9-12, 1981.

[190] Aaby T and Rosenberg B. Patent # 4344440 “Microprobe for monitoring biophysical phenomenon associated with cardiac and neural activity” USPTO granted 8/17/82.

[191] Holder DS. “Feasibility of developing a method of imaging neuronal activity in the human brain: a theoretical review” Medical and Biological Engineering & Computing 25(1): 2-11, 1987.

[193] Brodeur P.  The Zapping of America  Norton, New York, p 299 & 105, 1977.

[194] Berry S. “Decoding Minds, Foiling Adversaries” Signal Magazine 56(2): 56-7 Oct 2001. Also at http://cartome.org/brainmap.htm

[195] Pasternak D. “Reading your mind-and injecting smart thoughts” U S News and World Report p 67-8, Jan 3, 2000. Also at http://www.raven1.net/norsee2.htm

[196] Norseen JD. Mathematics, biofusion, and reflexive control for sentient machines. International Reflexive Control Symposium 17-19 Oct 2000, Russian Academy of Sciences, Institute for Psychology, Moscow, Ruaaia.  Available at http://www.acsa.net/norseen2000/

[197] Murray FJ. “NASA plans to read terrorists’ minds at airports” Washington Times, Aug 17, p A1, 2002.

[198] Lopez R. “Special operations survives Pentagon budget constraints” Jane’s International Defense Review 26(3): 247-51, 1993.

[199] McMurtrey JJ. “Microwave Bioeffect Congruence with Schizophrenia”  In press, 2004.  Available at  http://www.slavery.org.uk/MicrowaveCongruenceSchiz.doc

[200] Butler D. “Advances in neuroscience ‘may threaten human rights’” Nature 22 January 391: 316, 1998.  Also at http://raven1.net/nature1.htm

[201] McMurtrey JJ. “A Simulated Hallucination Mechanism Compared to Hallucination Brain Response Studies” 2007. [Online] [Cited 2008 Jan 3] Available from URL: http://www.slavery.org.uk/SimHallMechCompHallBrResStds.doc

 

[202] McMurtrey JJ. “Recording Microwave Hearing Effects: Literature Review and Case Report of an Affiant to Recording Remote Harassment” 2005. [Online] [Cited 2008 Jan 3] Available from URL: http://www.slavery.org.uk/RecordingMicrowaveHearingEffects.doc




Popular Posts - Last 30 days

Video: New Brain Computer Interface Technology - Steve Hoffman | TEDxCEIBS

The Matrix Deciphered - by Robert Duncan

Secret Testing - EM-Weapon Through Satellite

Mind Control: HAARP & The Future of Technology

U.S. Government Using Electronic Torture to Mimic Mental Illness

Voice to Skull Technology (V2K)

Understanding Vibration and Resonance

Neuropsychological & Electronic No-Touch Torture Report by Dr. Robert Duncan (used on targeted citizens here in the U.S.)

THE SEQUEL TO THE FALL OF THE CABAL | Parts 1 - 27

Bioeffects Research for Emerging RF Technologies