The AI Assessment Reality Check Educators Need
There’s a seductive idea floating around education circles: let AI handle the grunt work of assessment, question generation, difficulty calibration, and grading at scale. Sounds great in theory. The research landing this week tells a more complicated story.
A new study from Brazil put large language models to the test on something educators do instinctively: estimating how hard a question actually is. Researchers used ENEM, Brazil’s national university entrance exam, as their benchmark. The results? AI systematically underestimates question difficulty and struggles significantly with visual content.
Think about what that means in practice. An AI-generated assessment that claims to be “moderately challenging” might actually be too easy, giving students and instructors a false sense of mastery. Or worse, an AI reviewing your existing questions might confidently tell you they’re appropriately difficult when they’re not.
This isn’t a reason to abandon AI in assessment. It’s a reason to understand where the guardrails belong.
The Orchestration Model
Meanwhile, researchers at Carnegie Mellon developed ClassAid, a system that gives programming instructors real-time control over how AI assists students. The instructor can toggle between direct answers, hints only, or no AI support at all, adjusting on the fly based on what the class actually needs.
This is the model that works: AI as a tool that amplifies educator judgment, not a system that replaces it.
The ClassAid approach recognizes something important. Different students need different levels of scaffolding at different moments. A student who’s one insight away from a breakthrough needs a nudge, not an answer. A student who’s been stuck for 20 minutes might need more direct help. Only the instructor in the room can read that context.
Better AI, Still Human-Centered
The good news is that researchers are actively working on making AI evaluation more reliable. FairJudge, announced this week, specifically targets the bias and inconsistency problems that plague AI grading systems. It’s designed to provide more equitable evaluation across different assessment formats.
And a new approach to generating reasoning rubrics claims a 45% improvement in AI’s ability to identify errors in student problem-solving, particularly in technical subjects like math and chemistry.
These are meaningful advances. But notice what they’re optimizing for: making AI a better assistant to human evaluators, not replacing them entirely.
The Takeaway
If you’re exploring AI in assessment, go in with realistic expectations:
AI can generate questions, but you need to validate the difficulty yourself
AI can provide feedback, but students still need human judgment on complex work
AI can save time, but the time you save should go back into the teaching that matters
The Brazil study is a useful reality check. AI in education is genuinely powerful. It’s also genuinely limited. The institutions that thrive will be the ones that understand both.



