You can tell when a person is engaging in muscle activity by analyzing the squiggly lines of EEG readings obtained when someone puts on his head a device containing electrodes. Muscle movements of every type (including speech) produce deviations or disturbances in the wavy lines produced by EEG devices picking up brain waves. Because different types of visual images may produce different types of muscle movements (as illustrated in the visual below), it may be possible to predict above chance which of three photos a person is shown. Different photos may produce different types of muscle movements and different durations of muscle movements. So a computer program analyzing the squiggly lines of EEG readings may score above chance, by considering blips in EEG readings that may have different characteristics when different types of photos are shown. Such an ability is no evidence that brains produce minds, but merely evidence that different visual stimuli may produce different types of reactive muscle movements.
There is no brain-related technology that allows any person or computer program to figure out what a person is thinking by looking at MRI scans of a brain or EEG electrode readings of brain waves. But there are various tricks and cheats that can be used by someone trying to persuade you that he has decoded a person's thoughts or "inner speech" by analyzing brain states or brain waves. Below are some of these cheats and tricks.
Trick #1: The leveraging of failures of follow fast-paced hard-to-follow instructions. I have noticed this sleazy trick in some neuroscience papers. It is the trick of doing an experiment that requires an experimental subject to very rapidly switch between speaking a word and merely thinking of a word. So, for example, there may be a computer program that flashes instructions like this, with the instructions appearing on the screen for the times shown below:
Say "hippopotamus" (3 seconds)
Pause (2 seconds)
Say "asparagus" (3 seconds)
Pause (2 seconds)
Think "perfect" (2 seconds)
Pause (2 seconds)
Say "principle" (3 seconds)
Pause (2 seconds)
Say "asparagus" (3 seconds)
Pause (2 seconds)
Think "inventiveness" (3 seconds)
Pause (2 seconds)
When instructions like this appear on a computer screen, with a very fast pace, and rapid switches between the type of instruction, there is a good chance that a subject will sometimes fail to follow the instructions exactly. So during some percentage of the time that the subject was supposed to be only thinking of a word, the subject may be speaking in a word, in audible speech or all-but-silent speech or silent speech involving lip movement. This may allow a neuroscientist to brag about "above chance" results during intervals when supposed "inner speech" occurred. What is going on is that the instructions have been almost designed in a way to produce a fair amount of audible speech or all-but-silent speech or silent speech involving lip movement during intervals when subjects were supposed to be engaging in only mouth-motionless "inner speech."
Trick #2: Failing to prevent mouth-movement during intervals supposed to be "inner speech." There is a simple way to prevent or minimize muscle movement from the mouth during testing intervals that are supposed to be thought-only "inner speech." One way is to have a test subject wear something in his mouth designed to prevent any movement of the lips or tongue, with the subject wearing such a device during any test interval in which he is supposed to be engaging in speechless "inner speech." Another way is to make use of some specialized motion detector that will sound an alarm whenever the subject moves his lips or tongue. No such devices are used by neuroscientists doing experiments claiming to decode "inner speech." So whenever they claim that something involved only "inner speech" we should distrust such claims, and suspect that there was a lot of actual speech or muscle movement (audible or not) going on during the recorded periods of supposed "inner speech."
Trick #3: The word length cheat. I have noticed this sleazy trick in some neuroscience papers. It is the cheat of doing an experiment that attempts to predict which of a small set of words a person is thinking about, while leveraging the fact that some of the words have longer lengths than others. So, for example, in some quick-paced instructions appearing on a screen, a user may be asked to think (without speaking) one of these words: dog, chameleon, apple, hippopotamus, triangle. If the pace is fast enough, with enough tricky switches between "say this" and "think this," some little traces of muscle movement may show up in the EEG readings, even during intervals when the subject is only supposed to be speaking; and from the length of such muscle movement it may be rather easy to predict which word the user was asked to think of.
Trick #4: No exact specification of the experimental procedure. This is a very bad defect of most papers claiming to decode inner speech from brain scans or EEG readings. Such papers will typically offer some sketchy outline of the experiment that went on, without specifying the exact procedure. The rule of thumb we should follow is: regard as worthless any paper claiming successful experimental results which fails to specify in sufficient detail the exact experience that subjects underwent, in a way sufficient for someone to attempt a replication of the reported results.
Trick #5: Cherry-picking best results. Using multiple subjects and many different electrodes reading from different regions of the brain, a researcher can cherry-pick a best result from the many results (a result that might easily be obtainable by pure chance), and then try to give the impression that such a result was a typical result. Something similar would be going on if you had 20 people try to guess 50 five-digit numbers, and then had some visual graph heading bragging about "60% accuracy" with the fine print revealing that this was for guess target number 35 and guesser number 17 (when the target was 44392 and the guesser guessed 44291).
Trick #6: Leveraging data backdoors in a sneaky way. This trick goes on when some researcher claims that they got an impressive result "from brain scans" or "EEG readings" when brain scans or EEG readings were only part of the inputs used, with the success mainly coming from some data backdoor. An example is when researchers have subjects look at images obtained from the COCO image dataset. That dataset includes text annotations corresponding to each of the images, an example being that a picture of an apple may be labeled as "apple" or "fruit." So a computer program analyzing EEG readings while test subjects saw particular images can find out words corresponding to the observed image, by using the data backdoor of the text annotation corresponding to each image. With a little obfuscation and "clouding the waters," a success so unimpressive might be passed off as "mind reading" even though what is powering the success is 98% simply looking up the text annotations corresponding to the images, a feat no more impressive than looking up the definition of a word.
Trick #7: Leveraging sound inputs. Some people with speech problems have the ability to produce sounds when trying to speak, sounds that an average person is unable to understand. This may sound like someone trying to speak with his mouth filled with food. Some scientist may connect such a person to some EEG device, either one that is invasive (involving brain-implanted electrodes) or not invasive. Some computer program may then train on the person's speech while he is reading something or trying to read something. The computer may get a good idea about correlations between sounds that a human listener cannot understand, and words that a person is attempting to speak. Then the computer program may report success at "decoding" something that may be called "inner speech" or "brain states" or "brain outputs," even though the success is coming mainly from sound inputs rather than brain states. The effort may be wrongly called "brain-to-text" or a "decoding of brain speech" although such terms are inappropriate under such circumstances.
Trick #8: Leveraging phoneme or attempted phoneme EEG correlates. I noted before that muscle movements of every type (including speech) produce deviations or disturbances in the wavy lines produced by EEG devices picking up brain waves. There may be particular EEG correlates for particular phonemes or attempted phenomes that a person may make. So when someone makes the sound at the beginning of "achoo" and "apple," that may tend to produce a particular type of EEG blip; and when someone makes the sound in the middle of the words "cheese" and "sneeze," that may tend to produce some other type of EEG blip. So if you have a computer program that is trained to recognize such characteristic EEG blips, by training after someone connected to an EEG device tries to read some long body of text, that program may gain some ability to pick up lots of what a person is saying from his EEG brain wave readings. This may be described as "brain reading" although it is more accurately described as muscle movement EEG correlation reading. A program trained to recognize particular type of EEG correlates of phoneme pronunciation or attempted phoneme pronunciation may use some fancy AI "fill in the blanks" algorithm (possibly involving frequentist word-guessing or syllable guessing or phoneme guessing) to enhance some limited success it has at picking up EEG correlates of attempted syllable pronunciations. None of what I describe in this paragraph is correctly described as "decoding inner speech," although it may be described as that, particularly under some fast-pace hurry-up methodology in which a good deal of actual speech or attempted speech is occurring during two-second intervals in which someone is supposed to be only thinking of a word, because of a study design that almost guarantees there will be a large amount of this spillover "talking or trying to talk when you were supposed to only think."
Trick #9: The "as high as X percent accurate" trick. This trick is as old as the hills. You slice and dice the prediction results into something like 100 different portions, and pick the portion with the highest predictive accuracy. You then say something like "my method is up to 75% accurate," mentioning the accuracy of the most successful little portion, rather than the overall results.
Trick #10: Leveraging AI and large language models. An AI system that has trained on very many web pages and online books may be able to fill in lots of blanks in sentences, using guesswork based on word frequencies and the frequency of words used in a particular type of sentence or sentence fragment. So for example, if you have a fragment of a sentence such as "I'm hungry so __ ____ __ ____ ______," the AI system might be able to predict "I'm going to make some food" or some similar phrase as the missing part. Leveraging such AI systems, an experiment might produce some success level at "decoding inner speech" much higher than it would get without using such an AI system, particularly if some experiment uses carefully chosen test sentences of a type that allow an AI system to predict the full sentence from only half of the sentence.
The latest example of a misspeaking neuroscience paper boasting about decoding inner speech is the paper "Inner speech in motor cortex and implications for speech neuroprostheses" which you can read here. We get in the paper various boast soundbites that are not backed up by anything reported in the paper. The paper starts out by making the false claim that "Attempted, inner, and perceived speech have a shared representation in motor cortex." Speech is not represented in the cortex or any other part of the brain. The beginning of the paper contains quite a few untrue statements about the previous results of researchers, statements that are untrue because of various defects in the results published by such researchers.
Many of today's neuroscientists misspeak like crazy when they use the words "represent," "representations," "decode" and "decoding." Misstatements by neuroscientists using these words are extremely abundant. As a general rule you should never trust a neuroscientist using the words "represent," "representations," "decode" and "decoding." When it comes to "representations" neuroscientists are often guilty of very bad pareidolia and noise-mining, which involves a kind of seeing things that are not really there. Nowadays it easy for a scientist to kind of see things that are not there, by using "keep torturing the data until it confesses" tactics that often involve shady manipulations of data by dubious custom-written computer programs. We also should have a default distrust over any neuroscientist statement made by a neuroscientist about a decoding percentage accuracy. Such statements are typically extremely dubious, involving very dubious or easy-to-discredit calculation methods, or claims in which no calculation method is ever adequately specified. Often in a paper some impressive "decoding accuracy" figure is stated, but never justified.
Our first reason for distrusting the "Inner speech in motor cortex and implications for speech neuroprostheses" paper comes when we read that it involved only four subjects. As a general rule, correlation-seeking neuroscience experiments have no value unless they use a study group size of at least 15 or 20 subjects; and usually the required study group size is much larger.
Another strong reason for distrusting the "Inner speech in motor cortex and implications for speech neuroprostheses" paper comes when we consider the endangerment-of-the-sickest procedure that its researchers engaged in. The study involves invasively inserting microelectrodes into the brains of four very sick patients. This was not done for any medical benefit for these patients. The very sick patients had diseases such as the muscle-wasting disease ALS, sometimes called Lou Gehrig's disease. The insertion of microelectrodes into brains involves very serious medical hazards, and when used on very sick patients it may worsen their difficulties. In this case the very sick patients were used as "experimental guinea pigs," without any medical benefits coming to them from the medical risks they were enduring.
Whenever such shady business is going on, we should all-the-more tend to distrust any statements made by the people engaging in the shady business. We should nowhere be giving "the benefit of the doubt" when such researchers make grand boasts, but demand the clearest evidence that such boasts are justified.
In the case of the paper "Inner speech in motor cortex and implications for speech neuroprostheses" no such clear evidence is given. The paper fails to give any very exact specification of the experimental procedures it followed. But from its Supplemental Information document we should have the strongest suspicion that some of the tricks listed above were used.
When asked to produce "inner speech," instructions were given that seem designed to produce muscle movement rather than pure thought. According to Table S1, the instructions were these:""
- "Imagine mouthing the word. Focus on what your mouth, tongue, lips, jaw and throat would be doing and how they would feel."
- "Imagine uttering the word aloud. Focus on the sound you would be producing."
- "Imagine hearing me (or someone’s voice you know well) say the word, focus on the sound of my (their) voice."
The same table tells us that instructions such as these were alternated with instructions like these:
- "Say the word aloud (to the best of your ability.'
- "Mouth the word as if you were mouthing to someone across a room, without sound."
No comments:
Post a Comment