Skip to main content

The context you need, when you need it

When news breaks, you need to understand what actually matters — and what to do about it. At Vox, our mission to help you make sense of the world has never been more vital. But we can’t do it on our own.

We rely on readers like you to fund our journalism. Will you support our work and become a Vox Member today?

Join now

Machine-generated nonsense might score better than you on the GMAT

Robots, like this life-size, humanoid robot from the CeBIT computer technology trade fair, have gotten pretty advanced. But can robo-graders effectively evaluate humans’ writing?
Robots, like this life-size, humanoid robot from the CeBIT computer technology trade fair, have gotten pretty advanced. But can robo-graders effectively evaluate humans’ writing?
Robots, like this life-size, humanoid robot from the CeBIT computer technology trade fair, have gotten pretty advanced. But can robo-graders effectively evaluate humans’ writing?
John MacDougall/AFP via Getty Images
Libby Nelson
Libby Nelson was Vox’s editorial director, politics and policy, leading coverage of how government action and inaction shape American life. Libby has more than a decade of policy journalism experience, including at Inside Higher Ed and Politico. She joined Vox in 2014.

A computer-generated essay — which included the line “humankind will always subjugate privateness” — scored 5.4 points out of 6 when graded by the same software used to score essays for the Graduate Management Admission Test.

The writing program, created by the MIT’s former director of undergraduate writing, Les Perelman, can create an essay in under a second given up to three keywords. It’s meant as a theatrical critique of grading essays by machine, writes Steve Kolowich in the Chronicle of Higher Education:

Mr. Perelman’s fundamental problem with essay-grading automatons, he explains, is that they “are not measuring any of the real constructs that have to do with writing.” They cannot read meaning, and they cannot check facts. More to the point, they cannot tell gibberish from lucid writing.

Humans still grade GMAT essays, which have one human scorer and one computer. When the human and computer disagree — as they should in the case of Perelman’s sample essay — an additional person is brought in to resolve the dispute.

But computer scoring is generally on the rise, helped along by a high-profile study (which Perelman disputes) showing that machines can grade writing about as well as humans.

One of the groups writing Common Core standardized tests hopes to use machines, sometimes called robo-graders, as well as humans, when scoring essays on the tests next year. The tests are currently being tried out in classrooms; one of the goals for the field test is calibrating the software so it can score students’ writing.

See More:

More in archives

archives
Ethics and Guidelines at Vox.comEthics and Guidelines at Vox.com
archives
By Vox Staff
Supreme Court
The Supreme Court will decide if the government can ban transgender health careThe Supreme Court will decide if the government can ban transgender health care
Supreme Court

Given the Court’s Republican supermajority, this case is unlikely to end well for trans people.

By Ian Millhiser
archives
On the MoneyOn the Money
archives

Learn about saving, spending, investing, and more in a monthly personal finance advice column written by Nicole Dieker.

By Vox Staff
archives
Total solar eclipse passes over USTotal solar eclipse passes over US
archives
By Vox Staff
archives
The 2024 Iowa caucusesThe 2024 Iowa caucuses
archives

The latest news, analysis, and explainers coming out of the GOP Iowa caucuses.

By Vox Staff
archives
The Big SqueezeThe Big Squeeze
archives

The economy’s stacked against us.

By Vox Staff