Note: Click on any word on this page to experience Interactive Orthography.

Paradigm Inertia In Reading Science and Policy – Part 3: Learning Disabled Science

Back to Part 2: A Warning Shot from the Bush Administration

Everyone we interviewed agreed: a significant component of the challenge of learning to read (English) is recognizing unfamiliar words fast enough to keep comprehension primed and flowing.

 
What most challenges the brain and causes the processing delays that “stutter” the flow of reading, is the time it takes to recognize unfamiliar words. More specifically, the time it takes to work through the confusing relationship between the letters and sounds within unfamiliar words.


 
Yet, despite the obvious, and the implicit consensus, the leading scientists and policy makers exhibited an aversion to examining the role of code confusion in reading difficulties. The following video (2008) describes some of the reasons for this.


This aversion to scientifically examining the role of code confusion in reading difficulties took many forms. Some of the most powerful leaders in reading science and policy saw the code through the mental models of an already literate adult (rather than a pre-literate child’s mind):

 
In a moment that perhaps most exemplifies the paradigm inertia, we asked Dr. Reid Lyon (referred to at the time as a the Czar of Reading in America): “so what we are saying, in effect, is that the majority of our children, to some degree, are having their lives all but fated by how well they learn to interface with an archaic technology. Watch his response:

 
In effect Dr. Lyon was saying: we can’t change the code so there is no reason to think about the role of code confusion in reading difficulties (except in terms of training teachers). The code and its confusions are an immutable fixture in thinking about reading science and policy. Therefore the problem is not the code but the the lousy job teachers are doing teaching it.

We mean no personal criticism here. Dr. Lyon is someone we greatly admire for his championing of children. His response illustrates how the learning of even the best of scientists can still be disabled by their paradigms. As one of the nation’s most influential scientist / policy makers, the way he conceptualized the challenge of learning to read directly influenced government policy and academic research and indirectly influenced virtually every school in the country.

Perhaps the most stunning example of the paradigm inertia that pervades reading science is revealed in this exchange with Dr. Keith Stanovich, one of the most influential cognitive scientists in the field of reading.

 
Again, we mean no personal criticism here. Dr. Stanovich is a brilliant and caring scientist who contributed vastly to the cognitive science of reading. Yet, here we have one of the most highly respected scientists in the field of reading saying in effect: We’ve been badly burned by “reading wars” and how the irregularity in the code was used to justify whole language. We don’t want to go anywhere near those confusions and risk opening up those wars again. Here a scientist is saying: We don’t want to be scientific and learn deeper into the underlying issues that are causing so many of our children to have life-maligning difficulty learning to read because it might endanger our phonics train agenda and reopen those painful wars.
 
Having lived through the reading wars and feeling passionately that they were fighting for the future of generations of children, its easy to be sympathetic to these leaders.  Yet, the moment we set aside an instructional agenda, as we had with the Children of the Code Project, and instead focus on understanding the challenge, the reason so many kids (nearly 2/3 today in 2019) do not read at the proficiency level, it becomes clear that it is not the teachers’ fault. It’s the code.

Though we systematically (even if mostly unintentionally) blame and shame kids, parents, teachers (as well as improficient adults) for their difficulties, the root of their difficulties – the underlying confusion – is in no way their fault. It’s the legacy effect of a series of historical accidents in the development of the English writing system itself (see “First Millennium Bug“).

Many notables including, Benjamin Franklin, Noah Webster, Melvile Dewey, Theodore Roosevelt, and Mark Twain, recognized that the code’s letter-sound confusion was at the root of reading and spelling difficulties. But despite their efforts and those of hundreds of others, centuries of attempts to change the alphabet or reform English spelling – to render their relationship more simply phonetic – failed. The central issue is inertia; any change to the alphabet or spelling would create a ‘before’ and ‘after’ disconnect in the continuity of written English and it would be a disturbance, nuisance, and expense to everyone literate in the system as it is (for more see COTC Thoughts about Orthographic Reform).

 
Because changing the code – changing the alphabet or spelling – has such intolerable consequences,
our conceptions of ‘teaching reading’ have been constrained to accepting the confusion as immutable and, consequently, to paradigms of reading teaching organized around training the brains of readers to deal with it. Phonics and whole language methods are both attempts to compensate for (work around), rather than directly address, the confusing correspondence between letters and sounds. (see also Alphaphon analogy)

 

Stay tuned… Part 4 of the series will be coming soon….

, , , , , , , ,

Trackbacks/Pingbacks

  1. Paradigm Inertia In Reading Science and Policy - Part 2: A Warning Shot from the Bush Administration - Learning Stewards - May 28, 2019

    […] Part 3: Learning Disabled Science […]

Leave a Reply

This site uses Akismet to reduce spam. Learn how your comment data is processed.