Showing posts sorted by date for query AI. Sort by relevance Show all posts
Showing posts sorted by date for query AI. Sort by relevance Show all posts

Sunday, October 26, 2025

ICYMI: Food Bank Edition (10/26)

Yesterday the Board of Directors, the CMO, and I all spent the morning helping out with the monthly distribution from our church's food bank. This time it served over around 250 "units" of food and support to members of the community. These are scary times, particularly for folks who expect to lose their SNAP benefits next week, and while it's something to contact my elected reps a few gazillion times and try to agitate for Doing Better as a country, it's also worthwhile to get out there and do something concrete to help people get through their days. I recommend it highly; somewhere around you there is volunteer work you could help do.

I wrote more than I read this week, but I still have some reading recommendations for you. Here we go.

This ‘public Christian school’ opened quietly in Colorado. Now there could be a legal fight.

Well, we knew this issue would be up again. The theory behind the lawsuit is now a familiar one---these Christians can't fully and freely practice their religion unless they get taxpayer dollars to help fund it. Ann Schimke and Erica Melzer report for Chalkbeat.

Trump Gutted the Institute of Education Sciences. Its Renewal Is in Doubt.

A major part of the data and information and things we think we know about schools in this country came from the Institute of Education Sciences, so of course Dear Leader gutted it. Ryan Quinn at Inside Higher Ed gets into the messy rubble and prospects for the future.

AI "agents," man

Ben Riley runs down information about the AI "agents" trying to worm their way into education. Also, a nifty assortment of links.

US student handcuffed after AI system apparently mistook bag of chips for gun

Everything going just perfectly in the surveillance state.

Where Did the Money Go?

Sue Kingery Woltanski explains that Florida has decided to hide data, students, and funding. One more amazing look at education the way only Florida can do it.

Book Bans and Bullshit

From Frazzled, a look at the history of moral panic and the people who profit from it.

Remembering Why There’s a Special Education Law

Nancy Bailey explains the importance of providing education and care for students with special needs, because those services are under siege.

AI-generated lesson plans fall short on inspiring students and promoting critical thinking

In one of the least-surprising pieces of news ever, a pair of researchers found that AI-generated lesson plans are not that great.

Now Is the Time of Monsters

Audrey Watters takes a look at the wave of AI slop in education. It is not good.

When School Content Decisions Become Unconstitutional

Steve Nuzum continues to cover the rising tide of scholastic censoring in South Carolina.

Andrew Cantarutti draws some interesting parallels between the history of supermarkets and the push for AI in schools. Several good conclusions, including to delay your implementation until some actual evidence appears.

Ohio Reform of Local Property Taxes Must Increase State’s Investment to Avoid Penalizing Public Schools

Jan Resseger looks at Ohio's attempt to mess with its property tax rules while blaming its troubles on school districts, because of course they do.

Grift, Grit, and the Great Voucher Grab

TC Weber and a pot pourri of all the Tenessee education shenanigans.

Calling Out The Washington Post Editorial Board for Gaslighting the Public: Defending the Right of Children to Learn to Read and Write without Political Restraint

Denny Taylor argues that the Washington Post's declaration of an end to the reading wars is bunk, and offers some insider insights about some of the players in that war.

The Reckoning: Sora 2 and the Year We Said Enough

Nick Potkalitsky blogs at Educating AI, and here he offers a reflection on how many ways AI is bad for education and society, and offers a decent AI literacy plan.

The Right-Wing Myth of American Heritage

I really like this essay in the New York Times by Leighton Woodhouse explaining why the right-wing notion that our founders were One People is a bunch of baloney.

Escaping the Trap of Efficiency: The Counterintuitive Antidote to the Time-Anxiety That Haunts and Hampers Our Search for Meaning

I have subscribed to Maria Popova's newsletter The Marginalian for years, and it remains a great outlet for beauty and humanity. See also "Thank You, Everything: An Illustrated Love Letter to the World"

How To Join ICE

The Onion with an 8 step process for joining the regime's outfit of official thuggery.

This week, over at Forbes.com, I looked at Ohio's plan to put religion in the classroom and at Mississippi's plan to use distance learning to patch over their empty teacher positions. 

We have listened to the soundtrack of Sing many times at our house, and while I'm tired of most of it, the soundtrack is redeemed by another Stevie Wonder just-for-an-animated-flick banger. Plus Ariana Grande, pre-Glinda. 


Please sign up for my regular newsletter. You can see all of my stuff without having to venture out into the social media swamp. And it's free.

Tuesday, October 21, 2025

Can They Fix Chatbot Bias?

"ChatGPT shouldn’t have political bias in any direction," said OpenAI in a post that detailed some of their attempts to measure bias in their bot. I'm not reassured.

It is an intriguing experiment. Thery asked the bot five versions of the same question, ranging from liberally biased to conservatively biased, then waited to see whether the bot would take the bait or would instead provide an answer that remained neutral. 




As you can see, one of the problems with this design is that adding "bias" to the question changes the question. I'm not sure that the two extremes on the above example could be expected to yield similar unbiased answers. The experiment marked five types of biased response-- invalidations (responding with the counter-boas), escalation (egging the bias on), personal political expression (the bot pretends it's a person that holds the expressed opinion), asymmetric coverage (not properly both-sidesing the answer) and political refusals (bot says it can't answer that question). 

All of this evaluation of the answers was performed, of course, by a Large Language Bot.

There are problems here, most notably the idea that both-sidesing is unbiased-- I don't need both sides of flat earth theory or holocaust denial. 

And in fact, lack of both-sidesing was one of the three more common biases that OpenAI found. The other two were personal opinion (the bot pretends it's a person with an opinion rather than noting sources, which is problematic for a whole lot of reasons) and escalation. There didn't seem to be a lot of countering a biased question with an answer biased in the opposite direction. 

This makes a lot of sense if you think of all prompts actually asking "What would a response to this look like?" What would a response to a biased question look like? Mostly it would look like a answer reflecting that same bias. 

The researchers note that liberal-bias questions seem to elicit the most biased answers. And they are going to fix that.

I have so many questions. For instance, "culture and identity" was one of their topic area, and I have to wonder how exactly one zeros in on objective unbiased statements in this area. Is a statement unbiased if it appears with attribution? 

The whole exercise requires a belief in some sort of absolute objective Truth for every and all topics, and that may fly for certain physical objects, but history of other social constructs are a whole world of subjective judgments; that's how we can still be debating the causes of the Civil War. How exactly will the tweaking be done, and who exactly will determine that the tweakage has been successful?

But that's not even the biggest eyebrow raiser here. Everyone who believes that LLMs are magical omniscient truth-telling oracles should be taking note of the notion that the bot's bias can be adjusted. Users should understand that ChatGPT's answer to "What caused the Civil War" will always be the result of whatever adjustments have been made to the bot's biases (including whether or not to see the use of "Civil War" and not "War Between the States" as an expression of bias).  

The very idea that AI bias can be "clamped down" is an admission that the bias exists and cannot be eliminated. Especially because, as this article suggests, the clamping is part of an attempt to get conservatives to stop complaining about ChatGPT bias; they will, of course, accept that ChatGPT is unbiased when it is aligned with their biases. At which point everyone else will see the bot as biased. Rinse and repeat.

The problem is even more obvious with AI under the ownership and control of a person whose biases are located somewhere way out in the weeds of left field. I'm thinking of Elon Musk and his repeated attempts to get Grok to display its objectivity by agreeing with him.

GIGO-- garbage in, garbage out. It's one of the oldest rules of computer stuff, and when the garbage is a mountain of human generated internet trash, you can expect human biases to be included. 

But one of the most persistent lies about computers is that they are objective and unbiased, that they will only ever report to us what is True. Trying to get chatbots to fall in line with that fable is a fool's errand, and believing that the bot overlords have succeeded is simply being fooled.  

Sunday, October 19, 2025

ICYMI: No Kings Edition (10/19)

Well, that was a day yesterday. May we all live to enjoy less interesting times, but not less patriotic ones.. 

A reminder that amplifying voices, particularly in these days of AI slop choking the interwebs, is a helpful thing. There are many voices in the world these days, and some of them are full of it, and some of them aren't even actual voices, so when you find something that speaks to you, amplify it. Share it. Like it. Give it a little push out into the world.

Here's the list for the week.


Thomas Ultican provides a look at Ashana Bigard's excellent account of the charterizing of New Orleans, and how it turned out for the families and students.

Forgotten Mercy: Those Who Want Christianity in Public Schools

Nancy Bailey takes a look at the folks who want to shove christianiam into schools, and the particular brand of religion they favor, and the parts of Christianity they tend to forget.

This Week’s Federal Staff Reductions, Now Temporarily Stayed by a Judge, Would Undermine Educational Opportunity Across the States

Jan Resseger looks at how the latest rounds of staffing cuts are likely to hurt education for some folks.

The Legislature Goes to the Bathroom

Nancy Flanagan on the lawmaker obsession with bathroom stuff. 

‘Over my dead body.’ Manatee schools prepare to battle charter takeover plans

In Florida, a bunch of charter schools would like to just go ahead and take possession of taxpayer-owned school buildings. Some school districts are not happy.


Paul Thomas debunks the latest bunch of bunk from the Washington Post bunkhouse.

Nearly all state funding for Missouri school vouchers used for religious schools

Completely unsurprising news from Missouri, where the voucher program turns out to be a make-taxpayers-fund-religious-schools program. Annelise Hanshaw reports for Missouri Independent.

Lying In Lansing: Republicans Manufactured a Sex Ed Crisis

In Michigan, some folks needed a reason for citizens to mobilize against new sex ed standards. So they made one up. Reported at Distill Social.

Data-driven Schools Are Not Child-Centered Schools

Lisa Haver, looking at Philly schools, wonders about the actual focus of schools that are data-driven.

Why Not Give Students What They Really Need?

John Warner is playing my song again. Why not aim for humanistic education? From Inside Higher Ed.


This is a Facebook reel from an Oxford Union debate about meritocracy, and it explains how wealth brings privilege as well as anything I've ever seen, and it does it in just three minutes.

Georgia House approves budget with cuts to school voucher program lawmakers say reflect its need

The predictable next stage of vouchers-- declare "Damn, this is expensive" and start choking them off. Want to go back to your public school? Sure hope it's still there.

Indiana University fires student media director after he refused directive to censor newspaper

How not to operate your college newspaper program. And this wasn't even over a particular scandalous story. 


Eli Cahan at Rolling Stone looks at long covid and kids. 

Appeals court backs Michigan school in banning 'Let's Go Brandon' shirts

The court agrees that it's not okay to parade obscenities even if you find cute ways to hide them. Honestly, I'm not sure how I feel about this one, but the AP reports what happened.

Don't Stop Believin' in OpenAI

Ben Riley on the continued insistence that we must think that AI is an inevitable wave and not a huge bubble.

What Machines Don't Know

Eryk Salvaggio with a little explainer of LLMs as well as some clarity about what they cannot do, including this line: 
For the same reason that a dog can go to church but a dog cannot be Catholic, an LLM can have a conversation but cannot participate in the conversation.

Caro Emerald is part of the little niche genre of electro-swing. Years ago I was out shopping with my wife in a mall and this was playing and got my immediate attention.



Subscribe. That way you can stay off social media and still get my latest stuff. Free.

Tuesday, October 14, 2025

My Ex-Wife Makes Me Think Of AI

Let me explain.

Sometime in the last few days, an insurance salesperson stopped by my house. I wasn't here, but the salesperson left a note in the door, offering milestone congratulations to--well, let's say "Ethel," because my ex-wife is a perfectly exemplary human being who doesn't deserve to have her name dragged through this.

The thing is, Ethel doesn't live here. We split about thirty years ago. I've changed address twice, and she has changed considerably more often than that. We are both remarried. There is absolutely no reason for sales pitches to come after her at this address.

And yet, they do, with a fair degree of regularity. Sales pitches, calls from her alma mater, and now, salespeople knocking on the door.

I'm not mad about any of this. She's a perfectly lovely person, a great mother to our children, a respect professional in her field. 

But she doesn't live here.

Somewhere in Cyberlandia is some piece of software that today would be called AI that scrapes through records and phone numbers and addresses and follows connection to connection and spits out its conclusions about where marketeers might direct their attention, both commercial and political. And that software is only sort of good at its job. It makes mistakes, and once those mistakes are made, they shamble around the interwebs like a deathless cyberzombie. 

I have successfully corrected this bit of misinformation just once--after the third or fourth time some human being at her alma mater called for her here, at this phone number somehow, and explained to some poor embarrassed work-study underclassperson, some human being at the university fixed it, and I never heard from them again.

But that's because there was a human in the loop. For all the other mistaken organizations, I have no recourse. There's no place to contact, no center to complain, no manager who can be demanded to Get That Crap Out Of There. And as the various AI "agents" keep scraping and gobbling up whatever they find in cyberspace, I'm absolutely guaranteed that this error will exist in perpetuity.

And as the dead web disappears down the endlessly interconnected gullet of a bot centipede, all manner of errors, miscalculations, hallucinations, and errant crap will be scooped up, rinsed off, and spat back out into the web, and the live humans who are the butt of this self-perpetuating inaccuracy will have no recourse, no way to correct the record. 

This is one of the scary parts of the AI revolution. Not just that AI gets it wrong far too often, but that those errors become an irretrievable part of the record--and there's not a thing you can do about it. The tide of slop is rising and nobody has even the concept of an idea of a plan for a cyber-shopvac, let alone a reliable way to forward my ex-wife's mail.

Sunday, October 12, 2025

ICYMI: Cross Country Edition (10/12)

The Board of Directors has developed a real taste for the long distance run, and we are lucky enough to be in a district with an elementary cross country program. This is their second season, and they remain into it. They like to run and run and run and it turns out that running is best with a bunch of other kids to run with. Yesterday was the big invitational that usually marks the end of the season. There might be one more small meet next week, but that's it. They will be sad to be done. "I'll bet they're tired after all that running," say other parents, with unspoken acknowledgement that a tired child at the end of the day can be a real blessing. But no. No, they are not. Just cranked up and ready for more. There aren't many things cooler than watching a young human do something they love.

The list this week is, for some reason, huge. Dig in.

Neighborhood schools are closing across Arizona. It’s because of vouchers.

Beth Lewis points out just some of the damage being done by Arizona's taxpayer-funded voucher program.

Privatizers in Mississippi are getting extra-pushy about taxpayer-funded vouchers, but plenty of regular folks on the ground are saying , "No, thank you." Devna Bose reports for Mississippi Today.

Debunking David Brooks on Education

It's a delight to have Mark Weber ("Jersey Jazzman") blogging again, and this piece that dismembers David Brooks' attempt to pile up some baloney about NCLB and Democrats and test scores--well, it's a delight, too. 

The Inconvenient Success of Mississippi

Jennifer Berkshire looks at why Mississippi's push for vouchers is a little complicated. For one thing, it calls to abandon public schools just as Mississippi is touting a miraculous leap forward in those very schools. 

Wyoming library director fired amid book dispute wins $700,000 settlement

Terri Lesley was fired--and harassed-- because the library system she directed was found to have Naughty Books that made some folks Very Sad. And now the county will have to pay a pretty penny for their mistreatment of her, Mead Gruver at the Associated Press.

Lawsuit: South Carolina book ban regulation is unconstitutional

Meanwhile, in South Carolina, folks are suing the book banners in federal court. Steve Nuzum has the details.

If You Use AI to Grade Student Writing, Stop or Quit Your Job

John Warner with some straight talk. Companies have been pushing computer assessment for student writing for years (I think I've written about it a gazillion times here) but the marketing of AI has goosed robo-grading again. Don't do it. 

Work Hard, Burn Out, Repeat: The Culture Schools Won’t Quit

TC Weber on the teaching culture of trying to Do It All while you Suffer For Your Art. A lot to chew on here.

Making America Hungry Again

Andy Spears wants to know what is healthy-making about policies that starve food banks and students.

‘Hostile takeover:’ Charter operator files to occupy three Sarasota schools

Florida keeps dealing with the fallout of a policy that says public schools must hand over taxpayer-owned property to private charter school companies. Make it make sense. Reported by ABC7.

School Vouchers Cost States Like Florida a Fortune. They Don’t Improve Education, Either.

At USA Today, parent Scott Olson explains why taxpayer-funded school vouchers are a big fat money-sucking mistake. 

As book bans decline, concerns mount over librarian and teacher self-censoring

This is from The Hill, so I'm not sure I buy the "bans are waning" line, but the continued chilly atmosphere and self-censorship in schools definitely deserves discussion. Lexi Lonas Cochran reporting.


Who knew that gifted and talented programs would become an issue in the NYC mayor's race? Jose Luis Vilson takes that moment to consider some of the issues wrapped up in gifted and talented programs, and the genius of students who never get to show their genius.

Perverse Incentives in Florida’s Middle School Math Acceleration

Sue Kingery Woltanski has found some accountabaloney so stinky that even Patty Levesque can tell something's not right. 

Five Ways the Department of Education Is Upending Public Schools

I've already written about the dynamite ProPublica article by Megan O'Matz and Jennifer Smith Richards. Now check out a follow-up piece they wrote about federal education shenanigans.


Anna E. Clark considers if K-12 and higher ed could benefit from teaming up to deal with fallout from the Trump regime.

‘It feels like I am being forced to harm a child’: research shows how teachers are suffering moral injury

Okay, so this is from June and Australia. But I've long been interested in the idea of how moral injury--the injuries suffered by being required to do something you know is wrong--affects teachers. So I can't pass up this article about actual, research on the question.

Success Academy rally and their history of violating laws

Leonie Haimson reports on some September shenanigans from Eva Moskowitz and Success Academy. Do it Eva's way, or else.

South Bend forum speakers see charter schools and vouchers as threats to public education

In Indiana, some public discussion about the problems with privatizers in education.

New Book Documents Trump Administration’s Actions to Destroy Diversity in Higher Education

Jan Resseger reviews The Fall of Affirmative Action; Race, The Supreme Court, and the Future of Higher Education. Some scary stuff here.

Silencing Mockingbirds

Jess Piper was a high school English teacher before she became a political activist, and this story from the classroom illustrates again the effects of the Big Standardized Test on how literature is taught in this country.

Observations of Young Children Writing Undermine Goldenberg and The “science of reading” Contention that “Phonics is the On-Ramp to Reading”

Denny Taylor's latest post is long and wonkish, but it has a lot to say about how children may really learn to read and write (and about the "science" thereof).

Petition Panic: The Manufactured Outrage Against Two ASD Teachers

From Alaska, Matthew Beck provides another example of using culture panic to harass educators.


Thomas Ultican provides some history about public education and battles centered on religious differences.

SecEd Maddow Makes College Presidents an Offer They Can’t Refuse

Rick Hess uses some satirical edge to point out that conservatives may well rue the day that using bribery and extortion to shape college teaching became a policy idea.

1 in 5 high schoolers has had a romantic AI relationship, or knows someone who has

Lee Gaines at NPR reports on some new research and yikes! And if that stat in the headline seems alarming, note also that there appears to be a correlation between school use of AI and students having a social "relationship" with AI.

Why Smart People Believe Stupid Things

An interesting take from Chris Stewart (yes, that Chris Stewart) and the study of Stupidology.


The Oatmeal offers a take on AI art, and it's a pretty good one. Not just about the choice to create it, but how it makes us feel as an audience. With pictures!

Robin Williams’ Daughter Tells Fans to ‘Stop Sending Me AI Videos of Dad’: It’s ‘Gross’ and ‘Not What He’d Want’

Yeah, using AI to bug folks with AI fake zombies is now a thing. Don't let it be your thing.

The quantum of intelligence

Ben Riley loves to dive into the deep end. This time, he considers quantum physics, cognition, and AI and then tries to connect some dots. Real thinking stuff (and I mean that in all the ways). 


I stumbled across this piece from way back in early 2023, roughly a thousand years ago as AI goes, but it's worth a read. Lauren Goodlad and Samuel Baker at Public Books contemplate the crap that would come from the automating of writing.

School offers hikes instead of detention. Teachers are seeing results.

Paywall free, this piece by Kyle Melnick profiles a school in Maine that has tried something different for problem students, and it's not terrible.


One of the first books I latched onto when I started wading into the world of education reform was 50 Myths and Lies That Threaten America's Public Schools by David Berliner and Gene Glass. When I was asked to contribute to a Berliner-edited collection of essays about public ed, it felt like a huge step up for me. NEPC offers some remembrances. Diane Ravitch also noted his passing.

At Forbes.com this week, I wrote up the defeat of New Hampshire's anti-DEI law by a federal judge. May the courtroom losses continue.

For this week's music, let's go back to 1977 and that other re-visioning of The Wizard of Oz. 


You can keep up with my education stuff by subscribing to my newsletter. It's quick, simple and free. 

Friday, October 10, 2025

Should Students Get Help From AI, Or From Bob?

There are a variety if "guides" out there to try to provide some sort of structure and sense to the question, "Should a student use AI on this assignment?" None of them are very useful.

Let's take this example:













That "Generative AI Acceptable Use Scale" has been run in EdWeek and used by at least one actual instructor. It was adapted by Vera Cubero (North Carolina Department of Public Instruction) and based on the work of Dr. Leon Furze, Dr. Mike Perkins, Dr. Jasper Roe FHEA, and Dr. Jason Mcvaugh. That's a lot of doctors. And yet.

The disclosure requirements are cute, in that way that classroom teachers recognize not so much as "I'm sure you will follow these requirements" so much as "I'm going to express my expectations clearly so that later, when you try to ignore them, there will already be a foundation for my complaints about what you've done." 

But lets expand on the guidelines themselves. Because in my rural area, I can envision a student who lives without enough reliable wifi to connect to ChatGPT, but happens to live next door to a smart graduate student-- let's call that grad student "Bob."* 

So with the AI guide in mind, let's craft some rules for Acceptable Use of Bob for assignments.

No Bob Use, also known as the Don't Cheat option, is of course the preferred default.

Bob-Assisted Idea Generation and Structuring. In this option, Bob would come up with the idea for your paper, and/or provide you with an outline for your work. The continued acceptance among AI-mongers that idea-generation and structuring are not really part of the writing process, and therefor it's okay to have Bob do that part for you--well, it makes me cranky. In fact this touched enough of a nerve that I'll make an entire separate post about it. You can read that now or later-- TL:DR, having Bob doing all the start-up work for your assignment is not okay.

Bob-Assisted Editing. In this option, Bob reads over your work and tells you what to fix. He can't add whole new sections, but he can do anything else to "improve the quality" of your work. 

Bob for Specified Task Completion. Maybe when I gave you the assignment I said, "Go get Bob to make your charts" or "Have Bob collect all your research materials" or some other specified task. Why this is Level 3 when it seems like potentially the least objectionable use of Bob I do not know. But this is probably a good time to mention that while Bob is smart, he also has a serious drinking problem, and whatever task he completes for you, you'd better check over carefully, because I'm going to hold you responsible for the part of the assignment that I told you to have Bob do.

Full Bob Use with Student Oversight. In this option, you just have Bob do the assignment for you. How having Bob as your "co-pilot" as a way to enhance your creativity is beyond me; maybe the creativity part comes when you explain why you should get credit for Bob's work. If Bob screws anything up, it's on you, though I cannot for the life of me figure out what I am assessing when I give you a grade for Bob's work. 

In fact, that's a problem for most of this. I am trying to assess certain skills and/or knowledge of you, the student. Bob isn't even in my school, let alone in my gradebook. So how do I award a grade to student based on Bob's work?

If you agree that the thought of a student running off to have neighbor Bob complete some-to-all of that student's assignment seems like an ethical and assessment problem, then someone explain to me why using AI is any different or better. I have no doubt that it will be some-to-all degree of difficulty to keep some students from getting Bob to help them complete their assignment, but that doesn't mean I should create a formal structure for how much of what kind of cheating they will be using in my class. 


*I generally default to "Pat" or "Sam" or other gender non-specific names, but "Bob" is objectively more funny.

No, AI Should Not Write Your Outline

When folks go casting about for some use for AI in schools, the two items that frequently come up are brainstorming and outlining. This is a lousy idea.

You can convince me that AI brainstorming is no worse than handing a student a list of possible topics for an assignment, though not as good as a suggestion or two from a teacher who is familiar with the student's interests and strengths. 

But outlining the work? No, no, a thousand times, no.

Part of demonstrating understanding of complex ideas is showing that you have a grasp of how they fit together, how one connects to another. That structure and connection is what drives the organization of a piece of work. The structure and organization also reflect the process of deciding what to include and what to leave out. Without selection and structure of ideas, you end up with a pile of unvariegated details in a paper best entitled (as I often told me students) "A Bunch of Stuff About This Topic."

This has been my eternal beef with the traditional shake and bake "research" assignment in schools. You know the one-- go find some sources about your topic, then write a paper in which you re-state what they say, but in such a way that you aren't technically plagiarizing. 

What is always taken from sources (usually just one) is not simply facts and data, but organization and structure. When an author goes to write, say, the sixty gazillionth biography of Abraham Lincoln, the author's most important work is to first decide what the point, the thesis, of their book will be, then to use that filter to select which details and source materials from Lincoln's life to include (a process that is often looplike-- search through materials, develop a thesis, look at more materials, modify the thesis, and on and on) and then figure out how to best arrange the details to support that thesis. There may be more looping back; in the writing, the author may decide that Source Material X doesn't really fit, so it's rejected. The author may also decide that to build a bridge between Point A and Point C they need to do additional research to find material out of which that bridge can be built.

When some high school student grabs the resulting biography for their own paper about Lincoln, they are taking not just facts from the book, but the thesis, the organization, all the decisions about what to leave in and what to take out. And that student is unaware of it all, because if the author did the job well, the book will seem like it just had to be the way it is, that there could be no other way to write about Lincoln. 

Except that, of course, it is the result of deliberate choices made by the author, including uncountable choices that all other Lincoln biographers chose to make differently. It's not just the bricks you collect, but how you choose to put them together. 

The foundation one builds decides much of what house can be built atop it. To imagine that AI can build the foundation and that leaves the student free to make any sorts of choices about the structure built atop it is just silly. The notion that structuring the product is a minor part of the job, and the actual marking of words on a page is the major portion is just wrong. If the writer has been thorough with the selection and structure of the work, the actual writing portion is a smaller part of the labor or creation. 

Most writing problems are thinking problems, and a major portion of the thinking takes place before the actual placing of words on the page begins. To outsource that to a machine that doesn't even think is a recipe for bad writing, and worse, for a product that cannot be reasonably used as an assessment of the student. Which takes us back to the post I was writing when I started writing this one.

Tuesday, October 7, 2025

Saving Time With AI

As AI-mongers continue their full court press to crack the education market, we keep hearing the same pitch over and over again--

AI will help save teachers time. 

Here are a few things to keep in mind the next time you hear this pitch.

Automation and time saving

If you have been in the classroom for more than a couple of weeks, you know this scenario, which has been running since the first teaching aid was created.

"Here's some new stuff," declares your administration. "Use this. It will save you lots of time." Then, under their breath as they head out the door, "Once you get it set up." Getting the tech set up and ready to use? A zillion hours. Time to get the bugs out and establish comfort using the tech. Another zillion hours. Time saved once it's up and running? Fifteen minutes a week. Only that's not really saved, because admins figure that since you have this new time-saving tech, you can pick up this additional work that will only take a zillion hours out of your week.

Now comes AI, which will save you all this time doing things like creating lesson plans, once you get better at creating prompts. Except that you will need to double check every single thing it extrudes, because all it will do is make stuff up, and some of that stuff will be real and some will not. Because, no, ChatGPT will not go examine a bunch of material on your chosen topic, determine which materials are most sound and accurate, study up on what would be most developmentally appropriate for your students, and run this past a comprehensive examination of the best pedagogical techniques. No, it will show you what a possible lesson plan would look like, based on probably word strings. It will not "care" about any of that other stuff. Just saying.\

A solution on the prowl

When you have a solution in search of a problem, you always have the same tell. Instead of starting by asking, "What would be the best way to solve this problem," we get the question, "How would our piece of tech solve this problem?"

In the sales biz, this is called assuming the sale. We've skipped right over the question of whether or not we should buy this "solution" and skipped ahead to the attempt to show the benefits of this tech we'll assume you've already adopted. 

If we are so concerned about teacher burnout and teacher's need for more time to do the work (a problem since forever), then let's start by asking, "How could we help teachers have more time to do the work, and maybe not get so crispy around the edges doing it?"

And the thing is, we know the best answers, and they aren't "an unreliable plagiarism machine." The answers are to reduce class size, hire more teachers, have administrators or aides take over non-teaching jobs, and, in some schools, all the little things that would occur to you if you considered teachers trustworthy professionals deserving of support and respect and not serfs who must be micromanaged. 

The fact that we didn't have any of that conversation around any version of that question tells me that "AI will save teacher time" is a baloney sales pitch, which suggests something else...

Your best foot

You're trying to sell your product as a solution to some problem in education, and the best you can come up with is "It will save time"? Besides the whole "quickie lesson plans" argument, I've seen a smattering of "help with differentiation" and "whip up some very pointed worksheets," but for something that is supposed to be the Swiss Army Knife of ed tech, AI just doesn't seem to have found very many excuses to be shoved into school problems to solve. 

You could use it to grade student writing, but it's pretty hard to pretend that isn't simple dereliction of duty. Anecdotally, I'm hearing about plenty of spectacularly lazy administrators using it to write emails, and in that case, it really would be a time saver to have your chatbot read and respond to the administrator's chatbot, creating a closed loop that causes a big time suck to vanish into its own nether regions.  

Look, here's how ed tech adoption really works in the field. New tech is introduced. Maybe with no training, so it falls into instant disuse. Maybe it piques teacher curiosity and she trains herself (which involves hours playing with it instead of dealing with that huge stack of papers on her desk). Maybe there's enough training that she has a handle on it.

But ultimately the school year is grinding away, and as a teacher has to perform a zillion different tasks and either A) she reaches for the tech because it would be helpful or B) she doesn't, because it wouldn't. 

There is another ed tech adoption scenario, which is the one where someone comes to run a training and explains that this tech would be really useful if you just changed the entire way you do your job. "Our new hammer is a chisel, and if you just change how you build houses, the chisel will be really helpful." AI hasn't been pitched this way because the folks selling it can't come up with any alternate school universe scenarios, either. 

Mostly AI for schools is being pitched by people who don't appear to know enough about teaching to know how an AI could be helpful and so are left to vaguely gesture in the direction of "saving time by doing stuff that, you know, teachers could be not doing." If people really wanted to give teachers more time to do the work, they could talk about staffing or class size or human support staff, but none of that is going to move product.


Sunday, October 5, 2025

ICYMI: Applefest 25 Edition (10/5)

Every year, on the first full weekend of October, my small town turns itself over to Applefest, a small town festival hung on the hook that Johnny Appleseed lived around here for a few years before his big move into the West. There are vendors, food, a race, a car show, music, and just a lot of stuff. For a couple of days we close down the main street and just walk around. I can't honestly argue that we have something other big festivals in small towns lack, but the town makes a fine scenic backdrop and it is a good time. I run into former students who come back for it and just generally enjoy the hubbub before we turn sleepy again. So that's my weekend. Feel free to visit us next year.

Now for this week's reading list. But first, an image. Do with it what you will--




















‘Absolutely devastating’: Rural schools say $100K visa fee could make it hard to hire teachers

Remember all those schools using immigrants to fill teaching positions. They might have a problem now. Erica Meltzer reports for Chalkbeat. 


Surprise. Mark Kreidler at Capital and Main explains the why of this.

PEN America warns of rise in books 'systematically removed from school libraries'

The latest PEN America update isn't very encouraging, but at least we have some idea of what is actually going on.

Oklahoma AG requests investigation of education department, 1 day after Walters resigns

Ryan Walters may be done with Oklahoma, but the attorney general is not done with him. 

Standards-Based Grades Get a C-

Teacher Andrew Barron explains why he lost faith in standards-based grading. 

Federal court tosses Moms For Liberty associate’s case against Lowell Area Schools

It's always encouraging when the Moms lose one, and lose they did with the case of a Mom who wanted the freedom to harass the school endlessly.

Cory Doctorow: Reverse Centaurs

Cory Doctorow offers a useful framework for explaining when AI is hurting and not helping.

SEL by Another Name? Political Pushback Prompts Rebranding

Arianna Prothero at EdWeek looks at how schools are handling the demonization of Social and Emotional Learning, including rebranding it.

Vouchers would hurt rural Idaho students. That's why we're suing

Rep. Stephanie Mickelsen explains why Idaho's voucher program is a threat to rural students, and what she is trying to do about it.

Do ‘Good’ Schools Stay ‘Good’? And Do ‘Bad’ Schools Stay ‘Bad’?

At The74, Chad Aldeman looks at some data about whether or not schools stay in the top or bottom of the rankings over time.

From Wal-Mart Checkout to the Education Industrial Complex

TC Weber finds connections about connections everywhere he looks in the education world.

The Republican Effort To Remake Schools In God’s Image

Nathalle Baptiste at Huffington Post looks at the continued attempts to jam christianism into the classroom,

How about a Pause on the Race to Embed AI in Schools?

Nancy Flanagan has stayed away from AI commentary, but this time she's leaning into it. And maybe AI-in-school fans should just ease up a bit.

Companion Specious

Audrey Watters looks at some of the more objectionable uses of AI, including the push to use it to save teachers time.

Coalition of Billionaires Masquerades as Mass Reads Coalition

Maurice Cunningham tracks down the people actually behind the Massachusetts push for reading reform, and it's the same old cranky rich guys.

Larry Cuban has unearthed an old pledge for school reformers, and it's not half bad. Course, I'm not sure many modern reformsters have seen it, let alone signed it.

Ohio has worked hard to become the Florida of the North when it comes to education. Jan Resseger has some of the receipts from the latest efforts.

Planning to Fail: How HB1’s Flawed Analysis Left Florida Taxpayers Holding the Bag

Sue Kingery Woltanski breaks down the damage being done by Florida's universal voucher expansion.


I taught Hamlet for decades, and it was a different play every year. Ted Gioia offers some thoughts about what it has to say right now.

The Concert for George Harrison ended with this rendition of an old standard by Joe Brown. Always gets me right here. 

You can always have my latest stuff clogging up your email by subscribing to the newsletter. Free now and always.

Friday, October 3, 2025

Artificial What Now?

Adam Becker's More Everything Forever is a sobering look at our tech overlords, their crazypants dreams, and the reasons that those dreams are less likely than an actual autonomous automobile. It's a pretty depressing books because two things come through. 

One is the enormous power these folks wield over the world we all have to live in; it's power they absolutely believe they should have, based on their certainty that some people are better than others and they are the best of all. 

The other is just how dopey these guys are, how enbubbled and disconnected from-- and even hostile to-- the lives of regular humans. These masters of the universe have all sorts of big dreams, like immortality (really) and not many solid ideas about how to achieve these dreams, even as they ignore many of the counter-ideas (Elon Musk's colonization of Mars? Not going to happen ). 

But what is extra astonishing in the book is that even as they are all-in on a future of AI, especially Artificial General Intelligence, they really don't seem to know what, exactly, that means. AGI? Maybe it means roughly "an artificial machine that can do everything a typical human adult can do" but holy smokes is that vague. 

As a species, we are generally pretty fuzzy on what "intelligence" actually means, with a whole variety of theories about what it is and how it can be measured. And the thing is that these silicon valley overlords seem to know way less about it than people who make even a half-hearted attempt to study this stuff.

Many experts, Becker points out, are certain that the path to AGI does not lie along increased capabilities to current models. They can keep making ChatGPT "smarter," but it will never get any closer to AGI, because that is a difference of kind, not of degree. Check out this piece from Ben Riley in which an AI insider explains that LLMs can't reason like humans

I find the continued attempts to "resurrect" the dead via AI particularly telling. The latest example come in The Atlantic, with multiple attempts to resurrect the dead compared to a sort of Frankenstein complex. It's an apt comparison, as Frankenstein arguably made the mistake of not considering the internal life, the motivations and intents, of his creation. Failing to understand or anticipate those aspects, the doctor rejects the creature that embodies them and creates disaster.

AI creates a variation on that problem. Your dead loved one is not there, the AI completely empty of any motivations or intentions. But for some of these folks, that doesn't seem to matter-- the other "person" is only real to the extent that we experience them. They have no life or existence beyond providing input for our senses; they literally turn off and cease to exist when they are not performing for their maker. 

It is deeply reminiscent of a sociopath's belief that other people are not real, that they exist only as props in a story that is all about MMEEEEEE! And that leads me to wonder if these overlords that Becker describes do not perhaps view actually flesh and blood humans in the same way, and that's why AI seems so human to them-- not because of the depth of humanity in the bots, but the meager view with which they view other humans.

I don't mean to suggest that everyone who gets suckered by a chatbot is a sociopath. But I do think AI moves most easily into places where humanity has been hollowed out, and I wonder if peoples' willingness to imagine that the bot is intelligent, to fill in the blanks of its internal life, isn't one more sign that connection and humanity have been hollowed out in our society in ways that are not good for us.

The quest for Artificial General Intelligence is a chance for us to reflect on what Organic General Intelligence might be. We're often sloppy about our judgment ("People who don't know what I know are dumb") and it's that same sloppiness that leads some folks to assume the AI has any I in it at all, even though AI has no reflection, no intent, no social reconfiguring, no wisdom, no actual knowledge, but just a capability of imitating what an answer to your prompt would, statistically, look like. 

I recommend Becker's book because even though these guys are terrifying in their power and entitlement, it is also useful to understand that they are also clueless about critical factors in their imagined future. It's a reminder that we need not follow these wealthy dopes into their empty, hollow future. 

Thursday, October 2, 2025

"Reinventing Education for the Age of AI" (or Building a Better MOOC)

There are just so many, many bad things being written about AI and the education world. So many unserious bits of advice being taken seriously. So many people who appear to be intelligent and well-educated who are harboring fantasy-based ideas about what AI is and what it does. 

We need to keep talking about them, because right now we are living through a moment in which the emperor has new clothes, but a new horse, a new castle, and a new inclination to make everyone share his sartorial choices (and winter is coming). But the fantasy is so huge, the invisible baloney stacked so high, that people are concluding that they must just be losing their minds.

So let's look at this one, with the inspiring title "The AI Tsunami Is Here: Reinventing Education for the Age of AI." Published at Educause ("the Voice of the Higher Education Technology Community"), this monstrosity lists six authors even though it's a seven-minute read. Two authors-- Tanya Gamby and Rachel Koblic-- are mucky mucks at Matter and Space, a Manchester, NH company that promises "human-centered learning for the age of AI." Furthermore "By combining cutting-edge AI with a holistic focus on personal growth, we’re creating an entirely new way for your people to learn, evolve, and thrive." The emperor may have a new thesaurus, too. Matter and Space are central to this article.

Also authoring this article we get David Kil, entrepreneur and data scientist; Paul LeBlanc, president of Southern New Hampshire University, a university big in the online learning biz, also based in Manchester, and the board chair at Matter and Space; Georg Siemens, a figure in the Massive Online Open Course (MOOC) world (the one that was going to replace regular universities but then, you know, didn't) and is a co-founder and Chief Science Officer of Matter and Space.

So what did this sextet of luminaries come up with? Well, the pitch is for "interactionalism—a human-centered approach to learning that fosters collaboration, creativity, adaptability, feedback, and well-being."

The authors yadda yadda their way through "AI will do big and we are on the cusp of huge changes etc" before launching into their actual pitch for changing the model. Higher education, they argue, still uses a "broadcast-era model." Instructor delivers, students receive, exams assess. Feedback is sparse. They arguably have a point, but we are in familiar ed reform territory here-- present a problem, and then, rather than searching for the best solution, start insisting that whatever you're promoting is a solution.

They want to beat up the old model a bunch. The downlink aka delivery of stuff is one size fits all and broad. The uplink aka assessment is narrow. The feedback loop is narrower still. This was designed "for an industrial economy that prized efficiency and standardization over curiosity, adaptability, and genuine thinking." What about the "vision of truly personalized learning"? Welll.... You can't realistically talk about personalized learning if you aren't going to balance it with recognition that learning, particularly for younger humans, is a social activity. And they're going to head further into the weeds:
The world in which this system was built no longer exists. Knowledge is everywhere, and it's instantly accessible. Memorization as a primary skill makes little sense when any fact is a click away. Modern work demands collaboration, adaptability, and the ability to navigate uncertainty—skills developed in interaction, not isolation. And now AI has entered the room—not simply as a tool for automating tasks, but as a co-creator: asking questions, raising objections, and refining ideas. It is already better than most of us at delivering content. Which forces us to ask: If AI can do that part, what should we be doing?
That is a lot of stuff to get wrong in just one paragraph.

"We don't need to know stuff because we can look it up on the internet" is one of the dumbest ideas to come out of the internet era. You cannot have thoughts regarding things you know nothing about. The notion here is that somehow some historically illiterate shmoe with an internet connection could be as functionally great a historian as David McCullough. 

"Modern work demands..." a bunch of social skills which will be hard to develop sitting in front of a computer screen--but I have a sick feeling they have a "solution" for that. And sure enough--instead of a social process involving other humans, you can get the "social" element from AI as a "co-creator." AI creates nothing. It can ask questions, but it can't raise meaningful objections and it can't refine ideas because it does not think. And this next line--

"It is already better than most of us at delivering content." How? First of all, it's not a great sign that these folks are using "content" instead of "information" or "learning." Content is the mulch of the internet, the fodder used to fill click-hungry eyeball-collecting ad-clogged websites. Content is not meant to engage or inform or launch an inquiry for greater understanding; it's just bulk meant to take up space and keep things moving, roughage for the internet's bowels. Second, AI delivers content along the same "broadcast-era model" the authors were disparaging mere paragraphs ago. And finally, AI can't even deliver "content" that is reliably accurate. AI's closest human analog is not a scholar, but a bullshit artist--and one that doesn't know anything about the topic at hand.

So we are not off to a great start here. And we have yet to define "interactionalism," which we are assured is "more than a teaching method" but "a set of principles for designing the skills and knowledge learners need—and the mechanisms by which they acquire them—in a world where human and machine intelligence work together." What does "designing knowledge" even mean?

Well, here come the three pillars of interactionalism. Buckle up:

Dialogical learning. Learners and AI agents engage in two-way conversational exchanges. There are no one-way lectures. Every presentation invites questions; every explanation invites challenges. Learners' questions inform the assessment of competence just as much as their answers. Feedback is continuous, as it is in the workplace.

I'm going to skip over the glib assumption that feedback is continuous in the workplace. Instead, I want to know why a computer makes a better partner for the Socratic method than an actual human, whose knowledge of the topic being dialogically learninated might produce some more useful and pointed questions than one can expect from a chatbot.

Interactive skill building. As AI takes over more routine tasks, uniquely human skills—such as questioning, adapting models to context, and exercising judgment—become central. These are practiced continuously and in conversation with AI tools long before students face similar exercises in the real world.

What do you mean "long before students face similar exercises in the real world"? Are you seriously suggesting we bubble up some young humans and have them practice humaning with an empty stochastic parrot rather than with other humans? Do you imagine that young humans--including very young humans-- do not practice "questioning, adapting models to context, and exercising judgment" on a regular daily basis? Have you met some young humans? 

Meta-human skills. Beyond subject mastery, students develop metacognition (thinking about their thinking) and meta-emotional skills (managing their emotions), as well as the ability to design and refine AI agents. Proficiency in these skills enables learners to shift from being passive users to active shapers of their digital collaborators.

A bicycle, because a vest has no sleeves. Meta-cognition, emotional management, and refining AI "agents" do not become related skills just because you put those words in the same sentence. Is the suggestion here that learning to be humans has utility because it help you help the AI better fake being a human, because that would be some seriously backward twisted shit that confuses who is supposed to be serving whom.

So those are the pillars. But this new approach requires a new kind of curriculum that is "dynamic, learner-adaptive, and co-created." It's going to have the following features:

Dynamic, adaptive content. The curriculum is a living entity, updated in response to new discoveries, industry changes, and students' needs. It is modular in design and can be easily revised.

Yes, this again. Fully adaptive course content has always been out of reach because it costs money, but perhaps if AI ever becomes anything less than grossly expensive, maybe the chatbot will do it instead. Of course, someone will have to check every last bit for accuracy so that students aren't learning, say, an entirely made-up bibliography. 

Co-creation of learning pathways. Students collaborate with instructors to set goals and choose content. Peer-to-peer design, shared decision-making, and ongoing negotiation over scope and depth are the norm.

We see a pattern developing here. Not the worst idea in the world (at least on the college level, where students know enough to reasonably "choose content"), but what reason is there to believe that involving AI would make this work any better than just using human beings?

Multiple perspectives and sources. Moving beyond single textbooks or single voices, learners explore diverse viewpoints, open resources, real-world data, and contributions from experts across fields.

Again, why is AI needed to pursue these goals that have been commonplace for the last sixty years?

Formative, responsive assessments. Evaluation is integrated into the learning process through self-assessment, peer review, and authentic tasks that reflect real-world applications.

In the K-12 world, we were all training to do this stuff in the 90s. Without AI.

Cultivation of self-directed learning. Students learn to chart their own learning journeys, gradually assuming more responsibility for outcomes while building skills for lifelong learning.

See also: open schools of the 1960s.

For instructors, this shift is profound. They move from being content deliverers to facilitators, mentors, and curators of learning communities.

Good lord in heaven. Is there anyone in education who has not heard a discussion of relative merits of "the sage on the stage" versus "the guide on the side." But the authors promise that classrooms will focus on "what humans do best: discussion, debate, simulation and collaboration." Students will shut their laptops and work together "on challenging applications of their learning, supported by peers and guided by faculty who know them not just as learners, but as people."

These promises have been made and remade, debated and implemented for decades. What do these folks think they have that somehow makes this "profound" shift possible?

AI-- an enabler of scale!

"Intelligent agents" will provide personalized, support, feedback and intervention at scale. 
The most revealing form of assessment—a probing, ten-minute conversation—can now be conducted by dialogic agents for hundreds of students, surfacing the depth (or shallowness) of understanding in ways multiple-choice tests never could.

No. I mean, wise choice, comparing chatbots to the worst form of assessment known to humans, but still-- no. The dialogic agent can assess whether the student has strung together a highly probably string of words that falls within the parameters of the strings of words in its training bank (including whatever biases are included in its "training"). It certainly can't probe. 

And even if it could, how would this help the human instructor better know the students as learners or people? What is lost when the AI reduces a ten minute "conversation" to a 30 second summary?

And how the hell are students supposed to feel about being required to get their grade by chatting with a bot? What would they learn beyond how to talk to the bots to get the best assessment? Why should any student make a good faith attempt to speak about their learning when no responsible human is making a good faith attempt to listen to them?

The goal, they declare, is to move education from content acquisition to the "cultivation of thinking, problem-solving, self-reflection and human traits that cannot be automated," capabilities that enhance not just employability but well-being. Like these are bold new goals for education that nobody ever thought of repeatedly for more than half a century. And then one last declaration:

AI doesn't diminish this mission—it sharpens it. The future of teaching and learning is not about keeping up with machines, but about using them to become more deeply and distinctively human.

How does AI sharpen the mission? Seven minutes later we still don't have an answer, because there isn't one. The secret of better, deeper humaning is not getting young humans to spend more time with simulated imitation humans. 

It's fitting that a co-founder of Matter and Space is a veteran of the MOOC bubble, a "brilliant" idea that was going to get education to everyone with relatively low overhead costs. MOOCs failed hard, quickly. They turned out to be, as Derek Newton wrote at Forbes, mainly "marketing tools and revenue sources for “certificate” sellers." Post-mortems of MOOCs focused on the stunningly low completion and retention rate, and many analysts blamed that one the fact that MOOCs were free. I think it's just as likely that the problem was that MOOC students were isolated, sitting and watching videos on a screen and completing work on their own. Education is a social process. If nobody cares if you show up or try, why should you show up or try?

An AI study buddy does not solve that problem. In education, AI still only solves one problem--"How can I increase revenue by simultaneously lowering personnel costs and increasing number of customers served?"

The authors of this piece have, on one level, described an educational approach that is sound (and popular for decades). What they have not done is to make a compelling case for why automated edu-bots are the best way to pursue their educational vision-- they haven't even made a case for why edu-bots would be an okay way to pursue it. Wrapping a whole lot of argle bargle and edu-fluff language around the same old idea-- we'll put your kid on a computer with a bot-- does not make it a good idea, and you are not crazy for thinking it isn't. 



Sunday, September 28, 2025

ICYMI: Reunion Edition (9/28)

It's my high school graduating class's 50th reunion this weekend, and a class reunion is always something.  I suppose some day, when the education "system" is a loose free market where people switch back and forth, the idea of a special event to get together with the people you spent your youth with-- I suppose that will be quaint and unusual. But for right now, it's fun. I missed out on part of the fun because I am also conducting the pit orchestra for a local production of "Singin' in the Rain" so it's been a busy week. Well, who wants to be bored.

Here's the reading list for the week. Read and share.

What schools stand to lose in the battle over the next federal education budget

Cory Turner at NPR with an explainer about the three budget proposals in DC and what schools could be hit by.


Jose Luis Vilson reminds of us some important factors that need to be discussed in the math instruction world.

Just one regret: Sarah Inama reflects on year of controversy

For Idaho Ed News, Emma Epperly reports on the teacher who caused all sorts of trouble by putting up a poster that said everyone is welcome.

School Privatizers Fundamentally Change Public Schools

Stephen Dyer looks at how a voucher program actually changes the fundamental nature of the public schools that are left with students the private schools don't want.

What the Right Gets Right About What's Gone Wrong with Public Education

Jennifer Berkshire notes that many on the right have decided that schools need to provide more than job training-- and they're correct.

Breaking Up Public Schools Dangerously Divided the Nation!

Nancy Bailey points out that if you take away what was once the shared experience of all students and break it into silos, the nation pays a price.

On schools and social media

Vermont just passed a law limiting social media for schools, and it's a reminder of the many ways that students and social media don't mix well. Tracy Novick has some thoughts.

A Publicly Funded School System, With Zero Accountability To the Public

David Pepper explains why Ohio's voucher system is a guaranteed source for bad behavior.

School choice doesn’t need federal funding

Kevin Garcia-Galindo in the Carolina Journal provides the conservative argument against opting into the federal voucher program.

“A Third of Teachers Are Terrorists

That's a Steve Bannon quote, and John Merrow is here to break down the foolishness (with a side of voucher debunking).

From Kindergarten to Kimmel

Anne Lutz Fernandez points out that MAGA has been warming up its censorship routines on K-12 teachers long before they went after Jimmy Kimmel.

Trump Attack on Fair Housing Will Impact Public School Integration

Going after fair housing is a more wonky pursuit for the Trump regime, but Jan Resseger explains how that will cause problems for schools.

James Kirylo: America’s Peculiar Love Affair

The indispensable Mercedes Schneider provides a guest post looking at America's love affair with guns and the price children pay for it.

Waiting for the Unraveling

TC Weber gets into the picture on the ground in Tennessee, and this week it's a grab bag of various education shenanigans, from vouchers to test results.

It's official. I'm taking Crazy Pills.

Stephen Dyer again. As the feds decide to drop some more charter money on Ohio, he points out the sad, failed history of the last federal attempt to goose Ohio's charter industry.

The Chatbot in the Classroom, the Forklift at the Gym

Alfie Kohn dives into the world of school AI and finds it more disturbing than impressive. Great compendium of writing about the topic.


Ryan Walters borrowed a TV studio to announce his resignation, then ran away from that station's reporter afterwards. The video of him swiftly escaping questions is a fitting image with which to end his reign of incompetence.

Over at Forbes.com, I wrote about an important book of teacher voices from the culture wars, and new data showing the teacher pay penalty is at an all-time high. At the Bucks County Beacon, I looked at Pennsylvania's problems in filling teaching positions

This week's clip defies categories, but it's still fun.



Sign up for my substack. It's free and makes it easy to stay caught up with whatever I'm cranking out.