A week using AI teacher tools for lesson planning. Honest review

I spent the last week deliberately using AI teacher tools for lesson planning across three different units, something I’ve been resistant to because I’ve seen how quickly these things get adopted without much critical thought. I wanted to form my own view before saying anything publicly.

Here’s where I landed.

For generating a first-draft structure for a lesson, they’re genuinely useful. The kind of initial scaffolding that used to take 20 minutes of staring at a blank document now takes about five minutes of refining something the tool produced. That’s a real time saving and the output was surprisingly adaptable.

Where things got frustrating: the tools seem calibrated for a generic, compliance-ready version of teaching that doesn’t map well to how I actually run my classroom. The suggested activities are safe and competent. They’re also often boring. Anything that requires knowing your specific students, their reading levels, what they responded to last week, where the energy in the room actually is, that’s missing entirely.

The bigger concern is what happens when teachers who are newer or more overwhelmed start treating the tool’s output as a default rather than a starting point. The scaffolding stops being a draft and starts being the plan. That’s where I think the tool does real harm, not because it’s producing bad content, but because it’s replacing the judgment calls that good teaching actually runs on.

Tools don’t replace judgment. They can make it easier to find a starting point. But ‘here is a five-paragraph structure for teaching theme in short fiction’ is not a lesson. It’s a template that still requires someone who knows what they’re doing to turn it into teaching.

Worth discussing whether professional development around these tools is keeping up with adoption rates, because in my building it definitely isn’t.

The starting-point-versus-default distinction is the critical one. Any tool that positions itself as a drafting aid is fine. Any tool that positions itself as a replacement for professional judgment is a problem. The marketing language around these tools often conflates the two, which doesn’t help teachers or students make clear decisions about when and how to use them.

The calibration problem you’re naming is real. These tools optimize for something like ‘adequate and defensible,’ which is not the same as good teaching. They’re trained on examples of standard practice, not on examples of excellent practice, and that ceiling shows up quickly when you’re trying to design something that actually works for a specific group of students.

from the student side, the ‘safe and competent but boring’ description is accurate for like 80% of the worksheets I get. I always assumed that was just how school works but maybe there’s a reason now lol

The professional development gap you raise is the one that concerns me most in organizational contexts generally. Tools get adopted faster than the skills to use them well develop. That creates a period where the tool is being used but not well, and the outcomes suffer in ways that aren’t always attributed correctly to the tool gap.

as someone who TAs undergrads and interacts with a lot of lesson structures, the ‘template becoming the plan’ problem you’re describing also happens in higher ed. the tool-generated structure becomes normalized and then it’s very hard to tell whether you’re making pedagogical choices or just filling in boxes the tool gave you