Showing posts sorted by date for query AI. Sort by relevance Show all posts
Showing posts sorted by date for query AI. Sort by relevance Show all posts

Sunday, December 7, 2025

ICYMI: Chorus Edition (12/7)

The CMO of the Institute (that's Chief Marital Officer) sings in a community chorus because, among other reasons, she objectively has a voice much like that of an angel. It is one of those small town community things that we are able to enjoy, an example of how the people in a small town come together in a variety of different ways. Part of what is essential to the small town life is that you meet a lot of individuals in more than one context; that goes double for teachers. It's rare that you know a person just one way. The guys I used to buy my tires from grew up down the street from me and I used to babysit them; later I taught their children in school. And until they went out of business last month, I bought my tires from them, just like I used to buy them from their father. I have a million stories like that. Everyone in a small town does.

So anyway, this afternoon the CMO will sing with a community chorus an assortment of Christmas tunes and it will be lovely. December is an especially busy time for musicians. Please remember to show some appreciation to whatever musicians are lighting up your community.

Now for this week's list.

A Quiet Revolution Is Improving Schools

At The Progressive, Jeff Bryant points out that the community school movement is showing some real gains, even as the regime is not interested in supporting them.

How Trump 2.0 upended education research and statistics in one year

Jill Barshay at Hechinger Report takes us through the timeline of the regime's assault on actual education data and how that is taking us into a world in which we'll be flying blind when it comes to knowing what is really going on in schools.

Quinta Brunson wants thousands of Philly kids to have free school field trips

Philadelphia Inquirer has the story of Brunson's new field trip fund for Philly schools. Well done.

Kansas City tripled its share of Latino teachers in recent years so students can 'feel seen'

In 2019, just 1% of educators were Latino. The school district actually did something about it.

Which Parents Get “Parental Rights?” In This Ohio School, It’s Those Who Hate LGBTQ+ People

Parental rights are only for certain parents; that's always been a feature of the movement. Kelly Jensen at Bookriot shows how that plays out in one particular district.

Youth. For Christ? At School?

Youth For Christ is yet another group that believes the door is wide open for them to start recruiting in schools. Nancy Flanagan takes a look.

What is Heritage's "Phoenix Declaration"?

Several pieces have been written about this slice of baloney, but if you'd like one more look, here's a perspective from Steve Nuzum.

First Focus on Children’s Bruce Lesley Decries Trump’s Abandonment of Our Society’s Vulnerable Children

Jan Resseger looks into Bruce Lesley's dynamite piece about abandoning children.


Gary Rubinstein offers his review of Diane Ravitch's newest book. Do you have your copy yet? Get on it.

Teaching in a Season of Fascism

Matt Brady argues that when cruelty is policy, teachers are called to do some of the most important work in the country.

50 years after the birth of special education, some fear for its future under Trump

Happy birthday, Special Ed! Let's hope you've got a few more years left. Cory Turner looks at the occasion for NPR.


Cameron Dick's article will tell you nothing that you already know. But this piece was published at Zen Parent, and if you are looking for something to share with someone who is new to the issue, this is a fine choice.

3 states are challenging precedent against posting the Ten Commandments in public schools – cases that could land back at the Supreme Court

Charles Russo and Lydia Artz provide an overview of the current cases stacked up against posting the Ten Commandments in public schools.  At the Conversation.

Jeffrey Yass, Pennsylvania’s richest man, details how school vouchers drive his massive political spending operation in rare interview with Washington Post

The Washington Post ran an interview with Yass, but I don't have a WaPo subscription any more. However, the Philadelphia Inquirer ran a piece about the WaPo piece, and it captures many of the features of this rotten billionaire. 

Talking With Paul Kedrosky

Paul Krugman talks to Kedrosky and comes up with a pretty good explanation of AI stuff and why they are writing for a 37 year old guy on Reddit.

AI is Destroying the University and Learning Itself

If you want a really depressingly apocalyptic view of AI at the college level, Ronald Purser at Current Affairs has got you covered.

The People Outsourcing Their Thinking to AI

This Atlantic piece by Lila Shroff is worth it just for the coinage "LLeMmings." I love a good made word.

Flock Uses Overseas Gig Workers to Build Its Surveillance AI

From Wired, a reminder that sometimes AI isn't AI at all-- just a bunch of humans hiding behind a screen.

The New Anxiety of Our Time Is Now on TV

Ted Gioia connects Pluribus, fear, and the trouble with AI. 

Penn State Exceptionalism Meets Reality

What happens when a school (or, say, any organization) discovers that it's not really as exceptional as it likes to think it is? Full disclosure-- this is by Ben Jones, my nephew.

This week at Forbes.com, I took a look at two new challenges to the wall between church and state, part of the quest to feed even more tax dollars to private religious schools. Honestly, more people should have read this and paid attention, but at least a few months from now I can say, "I told you so."

This week's palate cleanser is Gregory Hines paying tribute to Gene Kelly


Feel free to sign up for my newsletter. 

Saturday, December 6, 2025

Reverse Centaurs, AI, and the Classroom

Cory Doctorow gave us "enshittification" to explain much of what has gone wrong, and he is already moving on to explain much of what we suspect is wrong with the push for AI. There's a book coming, but he has already laid out the basic themes in a presentation that he shared with his on-line audience. It doesn't address teaching and education directly, but the implications are unmistakable.

We start with the automation theory term "centaur." A centaur is a human being assisted by a machine. Doctorow cites as an example driving a car, or using autocomplete. "You're a human head carried around on a tireless robot body." 

A "reverse centaur" is a machine head on a human body, "a person who is serving as a squishy meat appendage for an uncaring machine." Here's his example, in all its painful clarity:
Like an Amazon delivery driver, who sits in a cabin surrounded by AI cameras, that monitor the driver's eyes and take points off if the driver looks in a proscribed direction, and monitors the driver's mouth because singing isn't allowed on the job, and rats the driver out to the boss if they don't make quota.

The driver is in that van because the van can't drive itself and can't get a parcel from the curb to your porch. The driver is a peripheral for a van, and the van drives the driver, at superhuman speed, demanding superhuman endurance. But the driver is human, so the van doesn't just use the driver. The van uses the driver up.

Doctorow explains that tech companies are highly motivated to appear to be growth industries, and then explains how they're selling AI as a growth story, and not a pretty one. AI is going to disrupt labor.  

The promise of AI – the promise AI companies make to investors – is that there will be AIs that can do your job, and when your boss fires you and replaces you with AI, he will keep half of your salary for himself, and give the other half to the AI company.

The thing is-- AI can't do your job. So the radiology department can't fire all the radiologists and replace them with AI to read scans-- they have to hire someone to sit and check the AI's work, to be the "human in the loop" whose job is to catch the rare-but-disastrous case where the AI screws up. 

That last radiologist is a reverse-centaur, and Doctorow cites Dan Davis' coinage for the specific type-- the Last Radiologist is an "accountability sink." Says Doctorow, "The radiologist's job isn't really to oversee the AI's work, it's to take the blame for the AI's mistakes."

In education, there is potential for AI to create centaurs and reverse centaurs, and I think the distinction is useful for parsing just how horrible a particular AI application can be. 

The most extreme version of a reverse centaur is any of the bullshit AI-driven charter or mini-schools, like the absurd Alpha school chain that promises two hours on a screen will give your child all the education they need. Just let the AI teach your child! All of these models offer a "school" that doesn't need teachers at all--just a "guide" or a "coach" there to be make sure nothing goes wrong, like an AI that offers instruction on white racial superiority or students who zone out entirely. The guide is a reverse centaur, an accountability sink whose function is to be responsible for everything the AI screws up, while allowing the investors in these businesses (and they are always businesses, usually run by business people and not educators) to save all sorts of costs on high-priced teachers by hiring a few low-cost guides.

For teachers, AI promises to make you a high-powered centaur. Let the AI write your lessons, correct your papers, design your teaching materials. Except that AI can't do any of those things very reliably, so the teacher ends up checking all of the AI's work to make sure it's accurate. Or at least they should, providing the human in the loop. So the teacher ends up as either a reverse centaur or, I suppose, a really incompetent reverse centaur who just passes along whatever mistakes the AI makes. 

Almost nobody is sales-arguing that AI can make teaching better, that an AI can reach students better than another human; virtually all arguments are centered on speed and efficiency and time-saving, and while that is appealing to teachers, who never have enough time for the work, the speed and efficiency argument is appealing to management because to them speed and efficiency mean fewer meat widgets to hire, and in a field where the main expense is personnel, that's appealing. 

Public schools don't have investors to make money from cutting teachers (though private and charter schools sure do), but for AI businesses (as with all other ed tech businesses before them) cannot help but salivate at just how huge the education market could be, a $6 billion mountain just waiting to be chewed up. So education gets an endless barrage of encouragements to join the AI revolution. Don't miss out! It's inevitable! It's shiny! To teachers, the promise that it will convert them into powerful cybernetic centaurs. To managers, the promise that it will convert teachers into more compliant and manageable reverse centaurs, controlled by a panel on the screen in your office. 

And both snookered, because an AI can't do a teacher's job. "Don't worry," the boosters say. "There will always be a human in the loop." Of course there will be--because AI can't do a teacher's job. The important question is whether the AI will serve the teachers or be served by them. As a teacher in the classroom being pushed to incorporate AI ("C'mon! It's so shiny!!"), you should be asking whether the tech will be empowering you and giving you new teacher arms of steel, or will it be converting you to some fleshy support for a piece of tech. 

Right now, far more pressure is being put on the Be A Fleshy Appendage side of the discussion. Here's hoping teachers find the strength to stand up to that pressure.

Oh, and a side point that I learned in Doctorow's article that's worth remembering the next time a company wants to offer AI-generated materials--  the courts have repeatedly ruled that AI-generated materials cannot be copyrighted (because they aren't human-made). 



Thursday, December 4, 2025

Glenn Beck's Patriotic AI Zombie

Well, something like this was inevitable.

The AI zombie market has been growing steadily. Schoolai caused a stir by unleashing an AI avatar of Anne Frank for classrooms as just one of their offerings of zombie historical figures for schools. In fact, there are now more outfits offering AI avatars for student use than I can even delve into here. Some are especially terrible; Wisdom of the Ages lets you chat (text only) with some big names of history, and within the first sentence, the Einstein avatar was talking about "he" rather than "I." Their "Adolph Hitler" also lapsed quickly into third person. Humy offers a Hello History app that promises all sorts of "engaging historical simulations" and an "in-depth and personal interaction with the historical figure of your choice." And don't forget the company that offers you the chance to take a writing class taught by a dead author. 

Then there's this horrifying ad from 2wai that promises to keep zombie Grandma around so that generations of your family can enjoy her. 



Good lord. And that's just one of many examples of the AI of Dead Relatives. I'm not sure what is worse-- the idea of dragging Grandma out of the grave or the idea that a few lines of code and some scanned letters and (2wai promises) a three minute conversation are all that's needed to capture a person's essence. No, actually, the worst part is that this encourages to understand that other people are only "real" to the extent that we perceive them and they reflect our expectations of them. These are simulations that amount to us speaking to our own reflections, empty images with no inner lives of their own. Simulacrums that exist only to provide us with an experience; voices that are silent except to speak to us. What the heck does that say about how we related to Grandma while she was alive?

Into this field of the damned comes Glenn Beck. 

Beck claims to have the "largest private collection of American founding documents in the world, surpassed only by the Library of Congress and the National Archives in Washington, D.C." And now Beck has plans for those documents, and they don't involve handing them over to a museum. Instead, on January 5, 2026, he will launch the Glenn and Tania Beck Foundation for American History, a privately funded trust, to make his collection of over a million documents accessible to everyone. 

It's the "next phase" of his career (post The Blaze), his "next disruption" and "creative venture." His foundation has created "the first independent, proprietary, AI-driven American historical library." It will come complete with its own AI zombie librarian named George, "built from the writings of George Washington himself. The writings of the Founders. The thousands of sermons that they heard from their church pulpits. The books that they -- they read. And the principles they lived by."

George is going to teach you the Real Truth, Beck promises. In fact, he guarantees that his AI will generate everything without hallucination or bias, which you might think is absolutely impossible for an AI (because it is), but Beck assures us that George is "contained within a secure, isolated server, where every document is memorized verbatim." Is there any other way that documents are stored on a hard drive?
This is not ChatGPT. This is not Wikipedia. This is verified, factual, memorized, first source truth.

Beck says that George will teach the Constitution, the Federalist papers, the civics. Beck says this project "will change EVERYTHING about education." George will counteract all those lies your teacher taught you. It's a proprietary AI database that will permanently preserve "the physical evidence of America's soul." 

There are at least two possibilities. One is that George will be a Washington-lite AI zombie that will, in fact, hallucinate and spew bias just like any other AI because Beck doesn't know what he's talking about. The other is that George has taken an old version of Jeeves and slapped a tri-corner hat on him, and that this is just a digital library with a search function because Beck doesn't understand AI, but he knows that it's a hot marketing term right now.

At least three outfits claim to have worked on an AI Zombie George Washington (here, here, and here) and they are all pretty much baloney. It makes sense that AI hucksters are going to go after the low-hanging fruit of public domain persons for zombiefication, and it makes sense that Beck, a seasoned patriotic grifter, would follow that path.

But boy is this shit a bummer, because Beck is going to wave his Giant Library around and convince a bunch of suckers that he can tell them the Real Truth about our nation's founders with even more unearned authority than he already deploys. But if AI zombies are good for anything, it's grift, and we had better steel ourselves for more of it. And please, God, keep it out of our children's classrooms.


Sunday, November 30, 2025

ICYMI: Pops Concert Version (11/30)


I play in a 169-year-old town band, and the day after Thanksgiving we present one of our biggest concerts of the year. It's a huge treat for us and audiences seem to enjoy it as well. It is how I wrap up the Thanksgiving holiday, though we get an extra-long weekend because here in NW PA, tomorrow is a day off from school because it's the first day of deer season. Hope your celebrations, whatever form they may take, have been pleasant as well.

Here's your reading list for the week.


ChatGPT has a teacher version now, and it stinks, Carl Hendrick points out some of the more egregious flaws (beyond, you know, using a bot to do your job).

The Quiet Collapse of Information Access

The AI School Librarian blog takes a look at some issues around access to information. Kind of scary stuff here.

EdTech companies are lobbying their way into your kids' classroom. Who's vetting them?

Well, you already know the answer, but Lily Altavena at the Detroit Free Press looks at the details.

What the Success Academy Scandal Says About the Charter School Model

Yes, there was a scandal, again, as Eva Moskowitz was caught, again, requiring her staff and students to be taxpayer-paid lobbyists for her charter chain. Ismael Loera at The Fulcrum connects the dots to the bigger picture.

Gratitude and Canned Goods—Teaching Children to Care

Nancy Flanagan considers one of those holiday traditions-- trying to get students to care about other folks and then do something about it.

The Elimination of the Professional Status of America’s Helpers!

Nancy Bailey looks at the details of the latest Trumpian kneecapping of teachers and other helping professions. Who was deprofessionalized, and what will that mean?

What to Know About Trump’s Definition of Professional Degrees

Another take on the same issue, from Jessica Blake at Inside Higher Ed. The whole thing may be a little more complicated than your social media threads make it out to be.

New Plan to Decimate U.S. Dept. of Ed. Exposes Trump Administration’s Deficient Educational Vision

Jan Resseger provides an excellent collection of reactions to and comments on the Trump plan to gut the Department of Education

The Education Department’s Forgotten Antiracist Origins

This New York Times essay from Anthony Conwright explains the history behind what the Department of Education was for in the first place. 

Teachers are outing trans students thanks to state’s new “Don’t Say Gay” law

Here's how Texas's Don't Say Gay law works out on the ground, with trans students outed and deadnamed. Greg Owen at LGBTQ Nation reports, and it's not pretty.

Souderton residents say school board’s Thanksgiving Eve appointment is a ‘lame-duck power grab’

Many conservative school board majorities were canned in the last election, but some aren't going to let a little thing like the will of the voters stand in their way.

Colleges Are Preparing to Self-Lobotomize

"After three years of doing essentially nothing to address the rise of generative AI, colleges are now scrambling to do too much." Michael Clune explains the trouble in the Atlantic.

Relationships First: A Skeptic’s Look at AI in Schools

Sue Kingery Woltanski is the skeptic, and this post offers some practical resources and questions to consider.

On artificial time

Chatbots can't wait, because they can't quite detect the passage of time. Ben Riley with more useful tech insights.

The Radical Power of Gratitude to Rewire Your Brain and Life

Thom Hartmann on research that suggests gratitude is actually good for you. 

From blast!, the who that demonstrates just how much you can do using a marching band as your building blocks.




Sign up for my newsletter, and get my stuff clogging up your email! For free! 

Thursday, November 27, 2025

Warding Off Classroom AI

There's a lot out there from folks trying and suggesting and selling ways for teachers to put their fingers in the dike holding back the allegedly inevitable AI tide. 

But I think playing AI whack-a-mole with computerized detector bots and policies designed for the express purpose of curbing chatbot cheating are not the way to go. Simply forbidding it is as effective as was the banning of Cliff Notes or Wikipedia. Numerous bots claim they can catch other bots in action; I am unconvinced and too many students have been unfairly and incorrectly accused. Trying to chase the chatbots away is simply not going to work. More than that, it is not going to help students grasp an education.

Cheating has always had its roots in a few simple factors. Students believe that success in class will be either too hard or too time-consuming for them. Students believe the stakes are too high to take a chance on failure. And students do not have a sense of the actual point of education.

I usually explain The Point like this-- education is the work of helping young humans figure out how to be more fully their best selves while working out what it means to be fully human in the world. That's a big soup with a lot of ingredients (some academic and some not), and the required ingredients vary from person to person. 

Because it's human.

As I've now said many times, AI most easily rushes into places where humanity has already been hollowed out. And unfortunately, too often that includes certain classrooms.

We've had chances to work on this before. Nancy Flanagan (and many others) tried hard to bring some attention to using the pandemic to reset schools into something better than either tradition or reform had created. But everyone (especially those in the testing industry) wanted to get back to "normal," and so we passed up that opportunity to reconfigure education. And so now here we are, facing yet another "threat" that is only threatening because we have created a system that is exceptionally vulnerable to AI.

Modern ed reform, with its test-centric data-driven outcome-based approach has pushed us even further toward classrooms that are product-centered rather than human-centered. But if class is all about the product, then AI can produce those artifacts far faster and more easily than human students. 

Carlo Rotella, an English professor at Boston College, published a New York Times piece that argues for more humantity in the humanities. He writes:
An A.I.-resistant English course has three main elements: pen-and-paper and oral testing; teaching the process of writing rather than just assigning papers; and greater emphasis on what happens in the classroom. Such a course, which can’t be A.I.-proof because that would mean students do no writing or reading except under a teacher’s direct supervision, also obliges us to make the case to students that it’s in their self-interest to do their own work.

Yup. Those the same things that I used to make my high school English class cheat-resistant for decades. Writing in particular needs to be portrayed as a basic human activity a fundamental function with lifetime utility. 

In education, it's important to understand your foundational purpose. It is so easy in the classroom to get bogged down in the daily millions of nuts and bolts decisions about what exactly to do-- which worksheet, what assignment, how to score the essay, which questions to ask, how to divide up the 43 instructional minutes today. Planning the details of a unit is hard--but it gets much easier if you know why you are teaching the unit in the first place. What's the point? I hate to quote what can be empty admin-speak, but knowing your why really does help you figure out your what and how.

If you have your purpose and your values in place, then you can assess every possible pedagogical choice based on how it serves that central purpose. The same thing is true of AI. If you know what purposes you intend to accomplish, you are prepared to judge what AI can or cannot contribute to that purpose. And if your purpose is to help young humans grow into their own humanity, then the utility of this week's hot AI tool can be judged.

Ed tech has always been introduced to classrooms ass-backwards-- "Here's a piece of tech I want you to use, somehow, so go figure out how you can work it in" instead of "Think about the education problems you are trying to solve and let me know if you think this piece of tech would help with any of them." 

But I digress. The key to an AI-resistant classroom is not a batch of preventative rules. The answer is to create a classroom with such a thoroughly human context, values, and purpose that AI is required to either provide something useful for that context, or is left out because it doesn't serve a useful purpose. The big bonus has nothing to do with AI, and everything to do with a more deliberately human approach to educating young human beings. 

Wednesday, November 26, 2025

Reformster Nostalgia And New Old Mistakes

There's been a recent uptick in reformster nostalgia, a wistfulness among Ye Reformy Olde Garde for a rosy past when there was a bipartisan consensus surrounding swell reform ideas like the free market and testing and the free market and No Child Left Behind and school choice and testing (e.g. Arne Duncan op-ed).

Mike Petrilli (Fordham Institute) has been substacking and gathering an assortment of all the old players to comment of education issues, running the gamut from A to B on various education policy debate topics, and in connection with that had a conversation over at Ed Week with Rick Hess (American Enterprise Institute) under the headline "Can School Reform Be Bipartisan Again?" Which is a question that certainly makes some assumptions, but let's take a look at what's going on.

Petrilli's stated motivation is fine. For one, he notices that substack is emerging as a way for people to scratch their writing and reading itch without having to slog through a variety of social media (some of which have become extra sloggy), and he joins a large club there (I know because I attend all the meetings myself). He also misses "the early days of Twitter and blogging, when we had robust debates about policy, tactics, and direction." Also understandable, and he explains what happened:
Unfortunately, as social media became a cesspool and the reform movement fractured along ideological lines, those conversations became full of vitriol and then largely went silent.

Sure. The ed reform coalition has always been complicated. The spine back in the day was a combo of free marketeers. social engineers, and tech/data overlords. Then Trump was elected, and then the culture wars were launched. Point to the moment when Jay Greene left academic reformsterdom and went to the Heritage Foundation and started writing pieces like "Time for the School Choice Movement to Embrace the Culture War."

It's not just that the ed reform movement became infected with Culture Panic. It's that the Culture Panic crowd is, almost without exception, a bunch of very unserious people. 

Over the past decade-plus, I've come to understand that the reformster tent is large and contains many different ideas and motivations. The reformster crowd includes folks who have some core beliefs and values that I believe are fundamentally flawed and the way to conclusions that I deeply disagree with. But they are people that I can have a conversation with, who use and receive words like their purpose is to convey meaning and not as some sort of jousting tool. 

The culture panic crowd is not serious about any of it. They are veiled and obtuse, deliberately misunderstanding what is said to them and using words as tools to manipulate and lever their desired results. They aren't serious about choice or educational quality or anything other than acquiring a dominant cultural position and personal power. There have always been some culture panic types within the reform tent (e.g. Betsy DeVos), but for half a decade they have been large and loud within the movement. "Let's use choice to encourage embettering competition" was replaced with "Get those trans kids off the track team." One of those is wrong, and one of those is simply unserious. 

Petrilli points to what he calls "reform fatigue," the result of two or three decades of hard push by reformsters. He calls it society's tendency to want the pendulum to swing back to the middle. "Eventually, the public grew tired, and the opponents of reform became more motivated than we, its defenders." 

He and Hess also point to the argument that Bush-Obama school reform was "simplistic and self-righteous," and Petrilli acknowledges the self-righteous part. Without naming Duncan, he says

I cringe when some reformers return to that self-righteous language, especially versions of “We know what works, we just need the political will to do it.” It’s a lot more complicated than that.

Petrilli also gives the movement credit for getting "big things" right, like the idea that "The American education system, with its 14,000 districts, elected school boards, and entrenched teachers’ unions, is not going to improve without external pressure." And he points to "student achievement" growing during the 1990s and 2000s, by which he actually means test scores.

Well, I think he's off the mark here. Fatigue? Simplistic? No, the reason that reform flagged was because it didn't work. Focusing on high stakes testing didn't achieve much, and most of what it did achieve was to damage school systems in numerous ways, from the narrowing of the curriculum to teaching an entire generation that the point of education is a Big Standardized Test. That and it became evident that test scores were a boon to data-grabbing tech overlords and people who simply wanted a tool for dismantling public education. 

The premise of a necessary "external pressure" is also problematic. Petrilli suggests that the pressure can come from "top-down accountability or bottom-up market competition," but I don't believe either of those will do what he imagines they will. Top-down accountability guarantees policies that are mis-interpreted as they pass down through layers of bureaucracy and which result in a compliance culture in thrall to Campbell's Law. Market competition is a terrible fit for education (see Greene's Law-- the free market does not foster superior quality; the free market fosters superior marketing). One of the bizarre fundamentals of the reform movement is the notion that educators are not doing a better job because they have not been offered the optimum combination of bribes and/or threats. 

Petrilli and Hess do not confront one of the fundamental flaws of reform, which is the notion that the Big Standardized Test is a good and effective measure of educational achievement, as if the question of how to measure something as vast and variable as the effectiveness of education is all settled. When David Brooks says that Republican states are kicking the Democrats' butts in education, all he's doing is comparing scores on a single math and reading test. As a country we have repeated this so many times that it is accepted wisdom, but the Big Standardized Test is just an emperor behind the curtain with no clothes. Will raising this student's BS Test scores give the student a better, richer, fuller, happier life than they would have had with their old lower scores? There isn't a shred of evidence for that assertion, but in the meantime, we keep pretending that a single mediocre math and reading test tells us everything we need to know about education.

Petrilli makes a passing reference to how unions never liked "testing, and especially accountability" (he has maybe forgotten their full-throated, member-opposed embrace of Common Core), which is just a rage-making assertion, because teachers and their unions have never, ever been against accountability. What they have opposed is accountability based on junk that has no connection to the work they actually do. Let's not forget that test scores soaked in VAM sauce gave us accountability measures that fluctuated wildly or that had to be run through other mechanisms in order to "evaluate" teachers via students and subjects they didn't even teach. The "accountability" created under Bush-Obama involved an awful lot of making shit up. 

Did test scores go up for a while? Sure. I was there. They went up because we learned how to align the schools to the test. Not to the education-- to the test. 

Petrilli muses about the nature of the reformster coalition, like the old one with members on the "ideological left, including Education Trust and other civil rights organizations" and I must confess that I never saw much "left" in the reform coalition. Petrilli says maybe we'll get back to a world where the parties fight over the center and then business groups and civil rights groups will become involved, and maybe, though reform has had plenty of chance to demonstrate how it can lift up minorities and the poor and it, well, didn't do that. If "populism" stays big, Petrilli muses, maybe they'll have to get involved with parents' groups and alternative teacher organizations "like the one that Ryan Walters now runs."

Well, except that would take them right back to a tent full of unserious allies who are not on the left, but are further right than Ye Old Reformy Garde. 

I'm inclined to ignore the right-left thing when it comes to ed reform. I think it's more accurate to frame the sides as pro- and anti- public education, and pro-public education voices have always been in very short supply in the reform coalition. Instead, reform positions on public education range from "Let's rebuild everything" to "Let's dismantle it and sell the parts" to "Burn it all down." 

Petrilli's smartest bit comes at the end:

For the people in the trenches, I’d encourage them to remember that student learning depends on student effort. And whenever they face a big decision related to curriculum, instruction, discipline policy, grading, AI policy, or anything else bearing on the day-to-day realities of schools, they should ask themselves: Is this going to make it easier or harder for my teachers to motivate their students to work hard and thus to learn?

This is actually pretty good, and it points to my suggestion for the imaginary new revived ed reformster coalition.

Include some actual teachers. 

I get there is a challenge here. In the same way that policy wonks and bureaucrats don't have real on-the-ground knowledge of teaching, teachers don't have real on-the-ground knowledge of policy wonkage and promotion. But ed reform continually misses the viewpoint of the people who have to actually implement policy ideas. 

Ye Olde Reformy Garde has come a distance since the days when they were hugely dismissive of teachers. Many have caught on to the fact that maybe deliberately alienating the people who have to implement your policy ideas is a poor choice. Maybe, just maybe, they've deduced, most teachers are in the profession because they really want to do a good job, and not because they are lazy sinecure-seeking slackers. 

But reformsters still miss the actual aspect of how their ideas play out on the ground, and those insights could save everyone a great deal of time. 

And no-- all those education reform leaders who spent two years with Teach For America do not count. Two years is bupkis; a real teacher is barely clearing her career throat after two years. 

Would working teachers just defend the current system so fiercely that no reform could happen? Of course not-- walk into any school in the country and the teachers there could tell you ten things about their system that should be fixed. Would teachers support accountability? Of course-- if it were real and realistic. Teachers have a powerful desire to teach next door and downstream from other teachers who are doing a good job. 

Lord knows I have no nostalgia for the old days of reform, when every year brought new policies that, from my perspective, ranged from misguided all the way to ethically and educationally wrong. Neither am I nostalgic for the days before modern reform. Public education has always needed to improve, and it always will, because it is a human enterprise. 

It would be great to have a reformy movement based on asking the question "How can we make schools better," but way too much of the reformster movement has been about asking "How can we get free market activity injected into the public school system" with answers ranging from "inject market based school choice" all the way to "blow it all up." It has marked itself by and large as an anti-public school movement since the moment that the A Nation At Risk folks were told their report had to show that public schools were failing and we were subjected to decades of pounding into the "common knowledge" that American schools are failing. And if the reform movement wants to revive itself, I suggest they start by owning all of that. 

We could have school choice, if that was what we really wanted, and we could have it without the segregation effects, the inefficiency and wasting of taxpayer dollars, without the pockets of really terrible education, without the instability of bad amateur players, without, in short, all the effects we get by trying to create free market school choice (I've explained how elsewhere).  But the reformster movement has long seemed far more interested in the Free Market part than the Improving Education part. They have spent forty years explaining that public education is failing because that's the justification for going Free Market (and national standards and high stakes testing) and yet it turns out that none of those things have been particularly helpful at all.

I do sense a new trend in Ye Reformy Olde Garde, and it's there in Petrilli's last paragraph-- a focus on policies "bearing on the day-to-day realities of schools." It's a good choice which might yield some productive discussions, particularly if those discussions are expanded to include people beyond the A to B gamut, because I know where you can find about 3 or 4 million people who are familiar with those day-to-day realities. 

Sunday, November 23, 2025

Grok Hilarious Fiasco Is A Serious Reminder

Grok is the Elon Musk version of AI, a chatbot that is supposed to be less woke. But lately it has also been a hilarious Elon Musk fanboy that will always tout the awesomeness of its owner.

It has boasted that Musk is "among the top 10 minds in history, rivaling polymaths like da Vinci or Newton." Also, he would beat Mike Tyson in a fight. Bruce Lee, too. Given the choice between Musk, Peyton Manning and Ryan Leaf in an NFL draft, Grok said, "Elon Musk, without hesitation." Musk is lean and muscular with extensive martial arts training, says Grok. Given the choice between switching off Musk's brain and wiping out the entire nation of Slovakia, Grok would vaporize Slovakia.

And that's before we get to ruder stuff, like Grok's assertion that Musk could be the best in all of human history at drinking piss and performing blow jobs. 

People started goading Grok with prompts like those asking Grok to praise Musk's ideas that he didn't actually have (e.g. historical theory about England's break with the Catholic church). At this point it's such a popular game that I'm not even sure if other examples are real, like Musk could be raised from the dead faster and more efficiently than Jesus-- but it sure seems they could be. Musk's crew, for its part, has been scrubbing away the brown-nosing Grokspeak, and Musk himself posted on X that “Earlier today, Grok was unfortunately manipulated by adversarial prompting into saying absurdly positive things about me. For the record, I am a fat [expletive].”

This is all a reasonably hilarious reminder of how over-the-top lying can get, as well as a look at how tragic it can be coming from someone who has more than enough resources to feel that he is--and has--enough. 

But it is also a reminder that one of the oldest computer principles still applies to one of the newest advanced technologies--

GIGO. Garbage In, Garbage Out.

Decades ago we used to hear those words as reminder that you have to be careful about what you feed your program because otherwise it will give back junk. Nowadays, we have to confront another aspect of GIGO-- when the people in charge of the program feed it garbage on purpose.

It was enough of a signal when Musk announced that he would make Grok "less woke," but now we've got a demonstration of how deliberately unmoored a chatbot can be, and not just because it's in its nature to make shit up, but because it can be "adjusted" to make shit up to fit a particular bias. 

I am daily frustrated by people who have fallen into the notion that AI chatbots are somehow paragons of objectivity that scan the web for the best evidence, weigh it logically, and deliver an unadulterated evaluation of the some total of human knowledge on whatever you have asked. They aren't, and they don't, and, in fact, you can "adjust" them so that they will always tell you that a certain man-baby is the most awesome human being to walk the earth--literally. 

Understand that this particular Elon-centric "adjustment" is so over the top that almost nobody would mistake it for an objective True answer. But you must ask yourself-- how many more subtle and less obviously bonkers "adjustments" could be made to the program that would not be obvious to you at all?

Even if you are deliberately trying to create an unbiased program, you will fail, because your own biases are reflected in everything that you "understand" about the world and what is in it. The notion that you can build a machine out of your biased pieces of mental lumber and come up with a house that is perfectly square is silly (also, the idea that perfectly square is how to build a house is yet another bias).

On the other hand, if you would like to build an AI chatbot machine that was biased in your preferred direction--well, that is totally achievable. It might take some practice to build the bias in subtly enough that you don't get caught, but with a little practice on top of huge amounts of wealth and compliant underlings, I'll bet you can get there.

We've talked a lot about AI as plagiarism machines and bullshit extruders and cognitive automaters, but we should be sure to include high-powered lie generators on that same list. Because as long as humans are on one end of the machine accepting output as an objective representation of reality and powerful folks are on the other end scripting the objective reality they would prefer, we are looking at some toxic high-tech GIGO.

ICYMI: Health Care System Edition (11/23)

My 92-year-old mother has spent most of this week in the hospital, and as always when I encounter the health care universe, part of me wonders how the hell people who don't have A) decent insurance, B) relatively easy access to a health care facility, and C) someone who can spend days sitting with them in the room, keeping them company, and translating and advocating-- I mean, my mother has all of those things, and it's still not super-easy. What the hell hope do people without those resources have? What a screwed-up system we have in this country, and yet some people insist on defending it avidly (and some other people would like to change the education system to more closely resemble it). 

So it's been a week here and it's not over, and if the blog has seemed a little quiet, that's why. I love you all, but I love my mom more. But I still have some pieces for you to read. 

School voucher confidential: Yes, the other parents are talking about you

Austin Gelder and Elizabeth L. Cline at Arkansas Times get commentary from a bunch of actual Arkansas parents about the state's voucher program. Nice change of pace, that, and not nearly as snotty as the headline might lead you to believe.

Learning with AI falls short compared to old-fashioned web search

Shiri Melumad has done some actual research indicating that people get more knowledge from a Google search than they do from an AI summary. And isn't that a low bar to fail to clear.

Are States Equipped to Track Students’ Paths From Classroom to Career?

Evie Blad at EdWeek asks many questions about the cradle to career data pipeline-- but not the most important one which is "Should we do this?" Informative yet awful.

RIP Department of Education

Jennifer Berkshire explains how education policies will be handled by the Department of the Boss.

‘Selling off the Department of Education for parts’: The agency’s major overhaul faces fierce backlash

If you want some official reactions to the news, 19th News has them.

I’ve already seen the impact from Charlotte’s Border Patrol surge

Juston Parmenter writes an op-ed for the Charlotte Observer (yes, that Charlotte) about the effects of the border patrol incursion. (Spoiler alert: the effects are not good).

Wall Street Is Paywalling Your Kids’ Sports

From The Lever, by Luke Goldstein. Turns out private equity has found yet another turnip to squeeze. And it includes not allowing you take recordings of your own child playing the sport.

Ohio is passing a law about a school exam question - A strange story behind a testing fiasco

When the Big Standardized Test screws up, does it take the state legislature to fix it? Ohio is working on the question.


Thomas Ultican notices that Erik Hanushek is out making wacky predictions again. What he's saying, and why you can safely ignore him.

Uncredible! ASD Debunks AG Cox’s Hillsdale Allegations, Citing Bishop-Era Policy

Continued noise and kerfluffle from the far right over Hillsdale pamphlets handed out in Anchorage schools.


Second part of a Jan Resseger series. It includes a link to Part I if you missed that, which you should,

Open Enrollment/Predator Schools

Andru Volinsky explains the trouble unleashed in New Hampshire by a state supreme court decision that facilitates an ALEC open enrollment scheme.

What is going on in Florida?

A lot, and almost all of it is unprincipled, anti-public education, and ugly (but not all of it). Sue Kingery Woltanski has the rundown, including the part where someone wants all public schools converted to a classical education. Plus the part where the state voucher system made $270 million go missing.

In Florida school wars, are locals finally pushing back?

Well, we can hope. Column by John Hill in Tampa Bay Times.

State Spending on Public School Students Lowest since 1997

That's the year they started voucherizing education. Ohio continues to shaft public school students, and Stephen Dyer has the numbers.

Federal judge rules law requiring display of Ten Commandments in Texas classrooms unconstitutional

This really shouldn't be news, but here we are-- no, you can't inflict your own particular religion on all school students.

Tennessee parents sue to stop voucher program

Opening shots fired. We'll see where the courts land on this one.

AI Suckage Round-up

An awful lot of news related to the awfulness of AI and its unfitness for education. Here we go--

‘We could have asked ChatGPT’: students fight back over course taught by AI

As I've repeatedly argued, you can't expect students to feel as if they should make an honest human effort when the people in charge of the course won't

AI Companies Are Treating Their Workers Like Human Garbage, Which May Be a Sign of Things to Come for the Rest of Us

Indeed. Joe Wilkins at Futurism

A general understanding of the human

Ben Riley hits several points, including classroom tech.

OpenAI Blocks Toymaker After Its AI Teddy Bear Is Caught Telling Children Terrible Things

Frank Landymore at Futurism says that at least OpenAI knew enough to pull the plug on sex fetishj instructions for children.

The résumé is dying, and AI is holding the smoking gun

See also: college recommendation letters.

The Great AI Bubble

Carole Cadwaller, the woman who used a TED talk to call Sam Altman a data rapist, explains the AI bubble and the economic disaster it will unleash.

The more that people use AI, the more likely they are to overestimate their own abilities

Ther's now some nifty research suggesting that AI will make your Dunning-Kruger problem even worse. "ChatGPT explained it to me, so now I am a freaking expert!!"

I started putting a music video into each of these weekly roundups because these days we can surely use a reminder about some of the nice, even beautiful, things that we humans create beyond policy arguments and political detritus. These are pieces of music I like, some for ages, and some newly discovered. Recommendations are welcomed. This week, it's Zak Abel, a performer I know nothing about, but I do like his song.



As always, it would be delightful if you subscribed. Always free.

Tuesday, November 18, 2025

Is There A Writing Crisis


Writing and writing instruction are facing a critical moment, an unprecedented techno-crisis, we hear. The word is spreading, the headlines announcing alarm. The death of English class. The end of student writing.

Students will outsource their writing assignments to ChatGPT or some other Large Language Model, a stochastic parrot only too willing to stand in for the student. Teachers will never see an authentic student essay ever again. Teachers are doomed! Doooooomed!

But AI couldn't threaten the writing classroom if we hadn’t spent the last few decades preparing the way. In too many classrooms, writing instruction is a fragile house of cards that was always going to collapse under the first stiff wind, and deserve to do so. As author and writing teacher John Warner put it, ChatGPT won’t kill any assignments that didn’t deserve to die. As Shannon Vallor put it, “AI can devalue our humanity only because we already devalued it ourselves.”

Writing instruction has long been wandering into a dark alley; AI was simply waiting at the end of that alley to mug its misdirected victim.

One of the oldest challenges of being human is both simple and profound. We are each individually locked into our own fleshy container, filled with ideas and emotions and images and impressions and memories, things we’ve experienced and things we’ve learned that we want to share with other humans, and yet we have no easy way to do that. We cannot project our thoughts and experience into another brain directly.

We are mysterious creatures, wonderfully and terribly made, our minds and thoughts and feelings trapped inside our bodies, and we spend our whole lives trying to master the business of making ourselves intelligible to the other body-bound souls around us. So we create systems of symbols, first of sound and then, later, of marks on a page to represent those sounds and the symbols they represent.

In his book On Writing, Stephen King answers the question “What is writing” with a simple line.

“Telepathy, of course.” And later, “a meeting of the minds.”

We dream of telepathy, mind-reading, of magical or technological ways to bridge the gap. We talk. We sing. We dance. We draw. We create entire languages and then expand and refine them in an attempt to make them better signifiers of everything bouncing around inside of us. And we write.

It is one of our most miraculous achievements. By making those marks on a page, we can cast our thoughts, our ideas, our feelings out across space and time.

But—and this is a big but—language, especially in written form, is both the most common and the most mysterious of human activities. It is an imperfect tool for making thought visible. We have no means to measure its effectiveness and accuracy, and we spend countless hours trying to interpret through the noise and parse the wording attempts of our fellow humans. “What did she mean by that?” can take up hours, even days.

That’s a central problem of education, where teachers are charged with reading the minds of their students. We can never be 100% certain that a student is truly and accurately communicating what’s in their head, because the only measure we have of what’s in their head is their communication. On any given day, we might be assessing what they hold in their head, or we might be assessing how well they are communicating it.

In many disciplines, we have worked our way around that central challenge. I can use straightforward objective measures to check for basic recall—do you recognize the desired piece of information when you see it? I can use performative tasks to create mini-artifacts of learning—what are the answers to these math problems?

But written language is the most complex and deep of human activities, and as soon as we try to reduce it to a simple performance or require that students produce an easily-measured artifact, we lose the plot. Asking students to perform a stripped-down version of writing, to write so that we can assess them (or worse yet, that some machine can assess them), is not the same as asking them to write for real.

Writing is thinking made concrete and visible. But by focusing only on the concrete and visible artifact of writing, and not the origin and process of its creation, the thinking behind it, we open the door to many other methods—non-thinking methods—to produce that artifact. And that’s where all our troubles, including those exacerbated by chatbots, began.

We have been living in a golden age of bad writing instruction. Text bots like ChatGPT are simply the latest step in a long slow march toward mechanical, performative writing.

This may be the oldest sort of in-school writing, the kind of performative writing in which the student jumps through hoops to prove to the teacher that she can jump through hoops.

Performative essays typically come with plenty of instruction about format and very specific expectations; the teacher may, even unwittingly, be suggesting that she has in her mind a picture of what the ideal essay for this prompt would look like, and she will be judging your work based on how closely it matches that ideal

This mechanistic model feeds on templates, as if there is some Platonic ideal of The Essay and all the writer is really doing is swapping out select words and phrases to match their assigned topic, trying to show they can pierce the ineffable to grasp that ideal.

The genre of performative writing tasks was boosted to even greater prominence by standardized tests, which come with strict and narrow requirements for what the writing-flavored artifact is supposed to look like.

These performative artifacts are for the ease and convenience of teachers and for the people who try to evaluate what teachers do. It’s all well and good for scholars to debate at inconclusive length whether or not Hemmingway or Chopin or Morrison are great writers and thereby open discussion to what the characteristics of great writing might be, but the front office needs a grade for Chad by next Tuesday, and the state needs to know if Chad’s teacher is highly effective or not, and besides, Chad is no Hemmingway, so let’s stop talking about aspects of writing quality and decide if Chad has produced a high scoring artifact.

Or maybe Chad’s teacher is thinking, “How am I supposed to judge whether this is any good or not? I’m not sure I can identify high quality writing, but I can damned well tell whether or not Chad followed the directions.”

Performative writing artifacts are attractive because they can be scored by an algorithm. Scoring by algorithm pre-dates scoring by computer. In the early days of the Pennsylvania state writing assessment, essays were scored by tables full of actual English teachers. We had the benefit of our professional expertise, but we were also quickly trained to follow a rubric which was an algorithm designed to guide us to the correct score. Soon, the state determined that anyone could be trained to follow that rubric, and regular teachers were displaced by temps answering classified ads.

Nothing follows an algorithm like computer software. But it’s important to remember, as vendor after vendor touts a new AI software product that can score essays, that computers do not and can not read and understand language in any sense in which we ordinarily use those words. They can only examine the external qualities of the writing and compare those qualities to however many millions of samples they have “learned” from. They too can only judge how well Chad has followed some rules; they just happen to be able to store a gazillion more rules than Chad’s overstressed human teacher. As demonstrated by the work of Les Perlman and his team at MIT over the past decades, computers are unequipped to deal with any inventive or unusual language use, while simultaneously being ill equipped to notice unusual content (like Abraham Lincoln leading the US through World War II).

Teachers at my school were able to crack the code of scoring well on state writing assessments with just a few simple rules. Recycle the prompt. Fill up lots of space, even if you have to be redundant to do so. Do not worry about factual accuracy. Use some big words, even if you’re not sure you’ve used them correctly (I was always partial to “plethora”). And it works. Our students scored well every time.

The performative model has been nurtured and cultivated in schools. It’s a performance students put on for their teacher, and, during the last couple of decades, for whatever lightly trained humans or barely competent software will actually set eyes (or sensors) on it.

This was the model favored by David Coleman, architect of the ELA standards for Common Core, but he had not invented it. He just seems to have internalized it from his own education. Writing for school is about writing to satisfy the requirements of the teacher. The teacher has spent weeks putting ideas into the student’s head, and Coleman’s description of how to read can be boiled down basically as “read this literature as preparation for writing a paper about it for your professor.” In Coleman’s world reading and writing are a closed loop that circles around a classroom. Having absorbed the teacher’s instruction, the student’s job is now to present the material for assessment.

Students are admonished to stay within the four corners of the text. At no point does this closed loop intersect any other part of reality—not with how the student understands the world, not with a grasp of culture, not with how the author communicates ideas that might spur action or vision of new generations.

So we crank out a parade of students who are equipped with performative writing skills that are of no direct use in the world, no help in the quest to bridge the gulf between one human and the world with which they want to communicate. Some students may rebel or encounter a teacher who swims against the tide (there are plenty out there), or they may, on their own, beat their experience into something useful. But mostly students have been taught not to write, but to comply with the algorithm.

ChatGPT and other Large Language Models are a reckoning, because they can also follow the algorithm, and unlike students, they don’t have to learn to block off portions of their brain to do it. Teachers are freaking out over the possibility that students may now be able to pull up an undetectable piece of cheating, a fully competent essay generated in seconds by a computer algorithm. While some folks are asking the question, “How shall we ever detect computer-assisted fraud,” the better question to ask is, “If this essay can be polished off by a computer, why are we asking human beings to write it?”

ChatGPT understands nothing, has no feelings or insights to convey, is no smarter than the paper on which we print words. It is stupid, stupid as a rock. It does what it does simply by following an assortment of algorithms, and so we have to ask—

Have we been teaching students to write, or to follow algorithms, to perform the steps laid out for them?

The question should make us uncomfortable.

Writing as a performative algorithmic act is, like a ChatGPT essay, empty of any meaning. It is not meant to communicate (and when students try to use it that way, they are too often slapped down). And, unfortunately, it has spread beyond the walls of schools. Demagogues and marketing executives treat language as a tool to draw desired behavior out of others. A former student of mine started his career writing short puff advertorial blurbs, like three sentences about a light-up toilet seat. It wasn’t, he said, really writing. The internet has increased the demand for writing (now known as “content”) that is not meant to communicate between human beings, but to fill up space, to attract eyeballs and/or the search engine bots, to then hand over the eyeballs to advertising.

It is its own empty form, and I suppose that mastering the empty performative forms of student writing might prepare students for that work—except that of course that kind of artifact can now be extruded swiftly and easily by AI. Knowing how to write like a robot is no help when the robots finally show up to do the job for far less pay.

I’ve talked to teachers and writing gurus who believe that if we can teach students how to manage these empty forms, they will somehow later in life be able to fill the forms with meaning and purpose and personal communication, and perhaps there are a favored few that manage that trick. But the vast majority of students get stuck on the question of “Why are we doing this empty exercise?” And the answer is nothing deeper than, “It’s a trick you must perform for your teacher and the state to get out of here and move on to a stage in life when you won’t ever have to do it again.”

Why teach students a skill that is not the real thing in hopes that the resemblance will somehow lead them to discover the real thing later?

Communication is the point of writing for real. It’s the whole object, the reason to bother with any of it. To start with the empty performance is like buying a suit for your future spouse when you are ten years old. And now, on top of thinking “Well, this seems pointless,” students can add, “Besides, there’s an app for this.”

Learning to perform the algorithm is about learning to fake. Fake interest in the topic. Fake the voice of a person writing about the topic. Fake some kind of developed insight or understanding. When begging my students to write authentically, I would say, “Life is too short to sign your name to a lie.”

The performative essay is also about compliance. It is about how well the student (and increasingly, also the teacher) will comply with the instructions, the requirements, the algorithmic demands. We find ourselves in a culture that values students who are willing to subsume their own ideas, their own expression, their own human desire to be understood by other human beings. That’s the chilling layer beneath this debate; if you are willing to give up freedom of expression, what other freedoms can you be convinced to give up?

Writing ought to be for real. It ought to be exciting and deeply human, carrying the full flush of putting yourself out into the world, to hurl your thoughts out into the intra-human void in hopes of connecting with others, of being seen and heard.

With all that in mind, this book is not about a concrete specific program—a set of teacher and student algorithms to follow. Trying to tweak the algorithm into a more human form is not the answer. Not lipstick on a pig, but a synthetic flesh face on a robot frame.

I started the thinking that went into this book long before generative AI started to grab the world by the eyeballs. ChatGPT did not create any of the issues discussed here, but it brought them into focus and demanded that we start thinking more carefully about what writing instruction—and writing itself—is for.

The focus here is on establishing a culture of writing in a classroom, about establishing for the teacher and the students (and parents) a mindset that allows writing for real not just to happen, but to thrive, and to become a tool that young humans can carry out into their lives, a real, living, vital human thing. All writing humans can use this set of values for re-focusing on the work. If these values are the foundation of a writing program, the program, whatever specifics are built into it (even technological ones), will be solid. In teaching, as in writing, knowing what you want to do makes it much easier to see how to do it. The spotlight here is on the classroom because that is the one place where we address writing deliberately, but the need for humanity and intention in our writing extends well beyond the classroom walls.

A thousand years ago, my eighteen year old self wrote an essay about why he wanted to go into teaching—a belief that if people were better at reading and writing and communicating, the world would be a marginally better place. There isn’t a corner of our world that would not benefit from fully authentically human communication. Perhaps by reflecting on what happens in a writing classroom, we can find some help in dealing with writing in the larger world.

Still here? This summer I took some time to finally hammer out a version of the book about the teaching of writing that has been rattling about in my head for too many years. This is the first chapter from that work, which has been sitting and waiting for me to figure out what to do next. In the meantime, I want to share this small bit. 

Sunday, November 16, 2025

AI Is Coming To Evaluate You

The unending tide of AI used for stupid things just keeps on coming, and as widely predicted, the major accomplices are managers and employers, sucked in with promises  that AI will make their work faster and easier and less have-to-deal-with-humans-y. Take the ars technica piece "The résumé is dying, and AI is holding the smoking gun." This strikes me as a parallel to teacher letters of recommendation, which are about fifteen minutes away from being wiped out by a mountain of near-identical and completely useless AI-extruded letters.

So it's no surprise when Technological Horizons in Education Journal is happy to pass along a PR release from Edthena about a tech tool that will do some of your principal's job for him. 

Edthena, mind you, is a company straight from AI hell. They've been around peddling old tech types of teacher coaching (watch yourself on video!) They have all your favorite PD buzzwords-- High Impact Feedback!! Amplify Coaching Capacity!! Scale Effectively!! Some of their marketing language feels... careful. "Evidence from video feels objective" they say, without addressing whether or not it actually is. 

And they're an approved platform provider for edTPA.

So they are a perfect business for AI-ing teacher observations into a useless stupor. 

Meet Observation Copilot! Your principal can feed it a half page of loose notes about what he saw in your classroom, and Observation Pilot will pad it with a bunch of professional and framework-aligned bullshit until you have pages of mind-numbing argle bargle in mere seconds. (No kidding-- the "demo" is below). The program will even generate suggestions for the teacher to implement, including all the approved soulless jargon, though unfortunately it does not appear that the program generates a suggestion to the principal that he either do his damned job or get the hell out of the profession. 

And you know that this "tool" is only about five minutes away from the concept of letting a video-cam collect the "observation notes" and thereby reducing the human principal's contribution to zero.

Sadly, there are actual testimonials here, like Brent Perdue, principal at Jefferson Elementary in Spokane, Washington. Brent says, "Observation Copilot has been a true game changer for me. It took that piece of the wordsmithing, of having the language flow, where I could really go down and just put in the facts of what I'm seeing." 

Or Juliana Addi, a school principal in Hoboken, who says, "Observation Copilot has changed my teacher feedback process. The writing that goes into it, it just expedites that pace - much quicker." Because speed is the important thing.

I can't begin to express the rage I would feel if a principal used this plagiarism machine to flood my evaluation with mounds of bullshit. I can only hope that the teachers who are subjected to this admin-o-bot respond by having ChatGPT write their response, or perhaps sitting in the post-observation conference and asking, "So what exactly did you mean when you write [insert quote here]." They should definitely do this while holding their copy so that the principal cannot see where the quote comes from in the fake evaluation.

This is of a piece with one through line of the LLM-in-education attack, which is the assertion that the business of turning a rough idea into a coherent sentence is an unimportant technicality that can easily be outsourced to a bot without any loss to whatever task is being completed, because human expression is no big deal. Just imagine.

Abraham Lincoln: "ChatGPT, just write me something about how this war is important to democracy and stuff."

Ernest Hemmingway: "Give me something booklength about how the Great European War made a lot of people sad."

Martin Luther King, Jr.: "As long as I'm sitting in this Birmingham jail, can ChatGPT just whip up some stuff about ignoring bad laws?"

Me, several years ago: "ChatGPT, please whip up something about love and getting married and stuff."

Yeah, stringing together the actual words-- that scary "wordsmithing"-- isn't all that important. Just have the bot do it.

AI most easily moves into places where the humanity has already been hollowed out. If you are a principal looking at this and thinking it seems like a super great idea, at a bare minimum, I hope you sit and have a hard think about your concept of your job. But maybe you should just think about alternate careers, because this kind of disregard for the human teachers who work for you is truly, deeply discouraging.

This is a terrible idea. Teachers need support from actual humans, not pages of jargonated filler from a bot that knows nothing about actual teaching. Teachers need to work in buildings where lines of communication are open, not ones where communication comes from a bot and not a human. Teachers need suggestions and ideas that come from a knowledgeable educator, not bot scrapings from the bottom of the internet bird cage. Useful assessment is a conversation between teacher and administrator, but to have that, both parties have to show up personally. For a principal to use this kind of tool (because I'm sure there are more out there) is unethical and disrespectful.

This little toxic AI menace is current available free of charge, because of course it is. The charging money part comes later, after you're so used to this crutch that you'd really hate to give it up. But with a dollar price of $0.00, using this tool will carry a higher cost than a school can afford to pay. 




ICYMI: Local Donuts Edition (11/16)

Apparently it is now our routine; Saturday mornings the Board of Directors and I go uptown to a local donut shop. Like most local donut shops, they make donuts way better than commercial ones because they are local and the donuts are not designed to travel cross-country and sit in packaging for days and days. You may remember the lesson of Krispy Kreme donuts, a once much-sought-after delicacy that was bought up and stripped of everything that made them desirable in order to market them at scale. There's a lesson there somewhere. In the meantime, if you're in my neighborhood, grab some Clark's Donuts. And in your own neighborhood, support a local business.

Okay, let's get to the list for this week.

"I Hope I'm Alive to See the Rebuilding Begin"

Scholar and writer Josh Cowen interviews Diane Ravitch about her new memoir, a book that you really ought to read.

Rigid Federal Rules May Block Efforts by Dem. States to Redirect New Federal Vouchers for Pro-Public School Uses

Jan Resseger points out that rosy predictions designed to convince Democratic governors to sign up for federal vouchers are--well, just not rooted in reality. Sorry, Arne.

The So-Called Science of Reading’s New Focus on Babies

Nancy Bailey explains the latest bad idea from the start-them-at-birth crowd. 

"Understanding" Shouldn't Be Vague or Mysterious

Michael Pershan looks at the idea of conceptual understanding and how to build it in a classroom. He's mostly talking math, but the ideas that he lays out here work just as well for other content.

How Pa.’s largest public cyber charter school discourages public participation

Oliver Morrison and the folks at PennLive have been doing an absolutely top notch series about cyber charters in PA. This is one of the best, and if I've done it right, this is a gift link (which will expire in six days). 

The Myth of the Anti-American Teacher

Nancy Flanagan covered these survey results last week, but here's the view from the other side of the education debates, via Robert Pondiscio.

Colorado’s 2025 election hints at future of state’s public education

Colorado is another state where outsiders spent a mountain of money on education issues-- and lost. Paula Noonan breaks it down.

How Much Screen Time Is Your Child Getting at School? We Asked 350 Teachers.

The New York Times talked to some teachers about screen time in schools. Some interesting data here, particularly as leaders are trying to jam AI into every corner of education.

Girl, 13, expelled for hitting classmate who made a deepfake porn image of her, lawyers say

Speaking of technology problems in school. This story comes from Louisiana, where I think they have some catching up to do.


Thomas Ultican  looks at attempts in California to avoid more charter school fiascos, and the history of pro-fiasco rich folks. 

Sticker Shock: Hillsdale’s Constitution (Propaganda) Revealed

Anchorage Alaska school district got in a Libs of TikTok-powered flap over disclaimers on Hillsdale free pocket Constitutions. Mathew Beck explains why it was bunk, and as a bonus, takes us inside one of those "free" handouts for students from the right-wing college.

Absence Makes the Smart Go Wander

Nancy Flanagan wins headline of the year contest with this piece about school absences (including the differences between public and charter schools).

What is Your Theory of Change?

Steve Nuzum looks at theories of change (or the absence thereof) in the education world.

Larry Cuban asks some pointed questions about "good" teachers.

Nov. 6th Local Election Returns Show Ohio Voters Out of Sync with Legislature’s Attack on Public Schools

Jan Resseger looks at the many ways that the election showed Ohio's legislature out of step with the voters.

Every Child Known… But Are They Valued?

TC Weber continues to dissect the policy that Nashville leaders claim, and finds his way to this poster-worthy principle-- Leadership that avoids accountability always protects systems that avoid responsibility.

Fallout from shutdown could be lengthy, school leaders say

At EdSource, Zaidee Stavely and Lasherica Thornton explain why even though the shutdown may be over, the mess it made in schools is not.

School Privatizers Cost Public School Kids $1.6 billion, or a fully funded public school system

Stephen Dyer corrects some deliberately misleading math.

Maybe Don’t Talk to the New York Times About Zohran Mamdani

This piece from Peter Coviello, former chair of Africana studies at Bowdoin College, is a great piece of prose and a fun read that along the way offers a dissection of New York Times failing style. From Literary Hub.

Can you really talk to the dead using AI? We tried out ‘deathbots’ so you don’t have to

So creepy. So very creepy. For The Conversation, by Eve Nieto McAvoy and Jenny Kidd.

A Pair of Billionaire Preachers Built the Most Powerful Political Machine in Texas. That’s Just the Start.

An important part of understanding Why Stuff Happens the way it does in Texas is understanding these two guys-- Tim Dunn and Farris Wilks. Excellent profile from Ava Kofman at ProPublica.

Sports Analytics: How Data Informs Preparation and Play

Unabashed plug. My nephew is a sports writer and his wife is an assistant professor of statistics at Carleton College, and I think it is beyond cool that they are teaching a course about sports analytics next summer. It's a non-credit course, and I think if you're anywhere around Northfield, Minnesota and have a love for wonky sports stuff, you should find a way to get into this. 

This week at Forbes.com, I wrote about a Florida book ban court decision that seems neither smart nor promising, and about the new set of Pennsylvania laws that will now rein cyber charters in a bit

Jimmie Lunceford deserves to be more remembered than he is. A seminal big band leader and an influence on guys like Glenn Miller, but under-represented in the video record. These guys could really cook. 


I'm hoping to crack the 3,000 mark for subscribers some day. It's quick and simple to subscribe, it's always free, and it's the easiest way to get my stuff regularly.