New Jersey Adopts AI to Grade Student Writing Exams
- •New Jersey to deploy AI scoring for statewide writing assessments across grades 3-10 this spring.
- •Automated systems will evaluate essays using parameters trained on previous human-scored practice examinations.
- •Concerns rise following significant AI scoring discrepancies recently documented in Texas and Massachusetts school districts.
New Jersey is transitioning to an automated approach for its standardized testing framework, the New Jersey Student Learning Assessments-Adaptive. This system utilizes artificial intelligence to evaluate essays and short-form answers in English Language Arts for students in elementary through high school. While computerized grading has existed for years, this shift represents a massive expansion in how the state handles the academic output of 1.3 million students.
The scoring engine, managed by Cambium Assessment, operates on strict parameters rather than generative models like those found in popular chatbots. This design aims to avoid "hallucinations," or the tendency of AI to generate false information. To maintain accuracy, the state claims that "unusual" responses will be flagged for human review. This hybrid workflow is intended to balance machine speed with human judgment, though the implementation remains controversial among teacher unions.
Critics point to failures in other states as a cautionary tale. In Texas, districts found that nearly 44% of AI-scored tests received higher marks when reassessed by humans. These discrepancies raise critical questions about the reliability of automated systems in high-stakes environments where a student's academic standing hangs in the balance.