"Train Them Then Dissect Them" is a phrase we can use to describe a particular type of animal experiment often done by neuroscientists. The experiment will work like this:
(1) Some animals (typically mice) will be trained to learn something. For example, they may be trained with the Morris Water Maze test to be able to go to a submerged platform within a water-filled tank after they are placed in such a tank. Or they may be trained over several days to keep their balance on a rotating rod, using something called a rotarod.
(2) The animals will then be killed, and their bodies dissected, with the brain cut up into slices that can be microscopically examined.
(3) The experimenters will look for some tiny area in the brain where they can claim to see some difference between the brains of the trained animals and a control group of animals who were not trained. All kinds of things may be checked for.
(4) The paper will make some announcement that some tiny difference was found between the trained animals and the animals in the control group. The reported difference might be any of 1000 different things, such as the size of dendritic spines in some tiny spot or the number of dendritic spines in some other part, or the length of synapses in some spot, or the thickness of synapses in some other part. The paper will claim that evidence has been found of "learned-induced remodeling" of the brain or "learning-induced modification of the brain." Neuroscientists will boast that they found evidence of a brain storage of memories.
There are several reasons why these type of experiments are typically very bad examples of junk science. One reason is that you can always find hundreds of tiny little differences in two randomly chosen animals of the same species, when microscopically examining their dissected brains. So merely by showing that there is some brain difference, you do nothing at all to show that such a difference arose from the training or learning that occurred in the mice. The same difference might have existed before the learning or training occurred.
An example of a very poor-quality paper following this "train them then dissect them" technique is the paper "Learning-induced remodeling of inhibitory synapses in the motor cortex." Glaring defects in the paper include these:
(1) The authors failed to use adequate sample sizes, with study group sizes such as only 7 mice or only 4 mice.
(2) The paper makes no mention of using any kind of blinding protocol, something essential for a paper of this type to be taken seriously. Neither the word "blind" nor the word "blinding" appear in the paper. The tiny differences reported in structures can easily be explained as being caused by biased ratings or biased size estimations being made by non-blinded analysts who knew which mice were trained and which mice were not trained, and who had a motivation to estimate in a particular way, so that statistical significance could be reported. The very tiny blurry barely-visible not-very-distinct hard-to-measure things being judged for size are just the type of things where bias of motivated and non-blinded analysts could be a big factor.
(3) There was no pre-registration of the study committing the authors to make a small number of checks for a difference of only one specific type in only one or a few exact spots. The authors were apparently free to keep checking in a hundred different ways, until some tiny difference was found somewhere.
(5) When the trained mice were compared to untrained mice, the control group of untrained mice was way too small, consisting of only 4 mice (Figure 2B). 15 subjects per study group (including 15 controls) is the minimum for any study like this to be taken seriously, with the required subjects probably being greater. No mention is made of a sample size calculation, which would have revealed how inadequate the study group sizes were.
(6) We have graphs supposedly showing some tiny difference found somewhere, but from a quick peek at the graphs you won't even notice any difference.
Even if the paper had shown a difference much larger, it would not prove anything, because anyone microscopically examining two randomly selected brains of an animal will always be able to find little differences here and there. It is never clear or probable that such differences occurred because one set of mice got training that the others did not. No good evidence of brain-stored memories is ever produced by such studies. The people who do such junk science experiments are needlessly killing mice.
I can give a description of how an experiment of this type could be done so that it would meet at least some of the requirements of robust research.
(1) There would be adequate study group sizes, such as maybe 30 mice in the group to be trained, and 30 mice in the control group.
(2) The study would be pre-registered, so that there would be a commitment to gathering data and analyzing data in a specific, limited way. For example, the specification of the pre-registration document might state that exactly 25 microscope-readable slices would be taken from the same region of each mouse, such as the hippocampus or the motor cortex, and that the study would only analyze one parameter, such as the quantity of dendritic spines.
(3) Each slice would be put in an envelope that had a subject number, and an indication of whether the mouse had been trained or not.
(4) A simple computer programs would be written that would have two functions: (a) the ability to generate a 7-digit random number and the ability to store in a text file a supplied subject number, a "trained" indicator of either Y or N, and that generated 7-digit number; (b) the ability to retrieve that subject number and its "trained" indicator when supplied the 7-digit random number. This is an elementary programming task.
(5) The program would be used to generate random numbers that would be written on envelopes. Each envelope containing a slice of brain tissue and a subject number would be replaced with an envelope containing one of the random numbers generated by the program.
(6) You would now have a set of envelopes marked only with random 7-digit numbers that a human could not recognize. Such a set of envelopes could be shuffled, and given to microscope analysts. Such analysts would thereby be completely blind to whether the tissue slices belonged to mice that had been trained or mice that had not been trained. There would be no chance of some bias effect in which an analyst tended to analyze trained mice differently. The microscope-using analysts would look for differences in tissue, using only the limited hypothesis to be tested that was stated in the pre-registration document. So, for example, if that document said that only the thickness of synapses would be analyzed, then only that one thing would be analyzed.
(7) After the microscopic analysis had been completed, and an analysis report form put in each of the envelopes, the computer program could be used to retrieve the original subject numbers and training status corresponding to each envelope. So, for example, someone holding an envelope with a random number of 4477353 might type in that number to the computer program, and get a reply of "Subject #21, Trained" or "Subject #35, Not Trained," with the answer retrieved from the text file previously made by the program. The answer could be written on each envelope.
(8) Then the data could be tallied up to see whether there was any difference between the characteristics of the trained mice and the untrained mice.
(9) Since the experiment would strictly adhere to the protocol of the original pre-registration design document, there would be no chance that the final analysis would include fewer or more brain slices than specified in that document.
That would be a decent design for an experiment of this type. No design like this is followed by the vast majority of these "train them then dissect them" experiments. Typically such experiments make no use of blinding at all. Any difference in the reported characteristics can be explained by either pure chance variation or bias of the microscopic data analyst, motivated to report some difference.
A paper such as "Learning-induced remodeling of inhibitory synapses in the motor cortex" tries to suggest that learning of a motor skill is stored as changes in dendritic spines. There is a reason why such a hypothesis makes no sense. Dendritic spines are very unstable things.
No comments:
Post a Comment