LLMs Don't Grade Essays Like Humans — But Here's What They're Actually Good At (API Tutorial)

🔥 Hot Take

  • arXiv paper (Mar 2026): LLMs grade essays inconsistently vs humans — confirmed
  • LLMs over-score short essays, penalize long ones with minor grammar errors
  • Smart move: use LLMs for writing generation, not grading
  • Build an AI essay assistant with NexaAPI — 10 lines of code, fraction of a cent per call

arXiv Bombshell: LLMs Fail at Essay Grading

On March 24, 2026, researchers published a paper that's making waves in both academic and developer circles: "LLMs Do Not Grade Essays Like Humans". The study evaluated models from the GPT and Llama families on automated essay scoring (AES) in out-of-the-box settings — no fine-tuning, no task-specific prompting.

The finding: agreement between LLM scores and human scores remains relatively weak. LLMs tend to assign higher scores to short or underdeveloped essays, while penalizing longer essays that contain minor grammatical or spelling errors. The models follow internally coherent patterns — essays they praise score higher, essays they criticize score lower — but those patterns don't align with how human raters actually think.

This is surprising to many developers who assumed LLMs could replace human graders. The paper says: not yet, and not in this way.

What This Means for Developers

Here's the nuanced read: this doesn't mean LLMs are useless for education or writing tools. It means developers need to use them for the right tasks.

What LLMs ARE reliable for in writing contexts:

The paper itself notes: "LLMs produce feedback that is consistent with their grading and that they can be reliably used in supporting essay scoring." The key word issupporting — not replacing.

3 Developer Use Cases That Actually Work

1. Generate Essay Drafts for Training Datasets

If you're building an AES system, you need training data. LLMs can generate thousands of essay variations at different quality levels for a fraction of the cost of human writers. NexaAPI gives you access to GPT-4o, Claude, and open-source models through one API — pick the right model for each generation task.

2. Build Writing Assistance Tools (Not Graders)

The research confirms LLMs are good at generating feedback that's internally consistent. Use that strength: build tools that suggest improvements, flag weak arguments, or propose better transitions. Frame it as "AI writing coach," not "AI grader." Users trust it more, and the AI actually delivers.

3. Automated Content Generation for E-Learning Platforms

E-learning platforms need massive amounts of content: practice prompts, model answers, rubric explanations, study guides. LLMs excel here. At NexaAPI's pricing, you can process thousands of content generation requests for dollars, not hundreds.

Build an AI Essay Assistant in 10 Lines of Code

While LLM grading may be imperfect, using LLMs for text generation via API is extremely cost-effective. Here's how to build an AI writing coach using pip install nexaapi:

Python — AI Essay Writing Coach

from nexaapi import NexaAPI

client = NexaAPI(api_key='YOUR_API_KEY')

# Generate essay feedback and improvement suggestions
# Note: NOT grading — coaching and improving
response = client.chat.completions.create(
    model='gpt-4o',  # Check nexa-api.com for latest available models
    messages=[
        {
            'role': 'system',
            'content': 'You are a writing coach. Provide constructive feedback on essays, '
                       'focusing on structure, clarity, and argument strength. '
                       'Do not assign numeric grades.'
        },
        {
            'role': 'user',
            'content': 'Please review this essay introduction and suggest improvements: '
                       '[ESSAY TEXT HERE]'
        }
    ],
    max_tokens=500
)

print(response.choices[0].message.content)
# Cost: fraction of a cent per request via NexaAPI

JavaScript — AI Writing Feedback API

// Install: npm install nexaapi
import NexaAPI from 'nexaapi';

const client = new NexaAPI({ apiKey: 'YOUR_API_KEY' });

async function getEssayFeedback(essayText) {
  const response = await client.chat.completions.create({
    model: 'gpt-4o', // Check nexa-api.com for latest available models
    messages: [
      {
        role: 'system',
        content: 'You are a writing coach. Provide constructive feedback on essays, '
               + 'focusing on structure, clarity, and argument strength. '
               + 'Do not assign numeric grades.'
      },
      {
        role: 'user',
        content: `Please review this essay and suggest improvements: ${essayText}`
      }
    ],
    maxTokens: 500
  });

  return response.choices[0].message.content;
}

getEssayFeedback('Your essay text here...').then(console.log);
// npm install nexaapi — cheapest LLM API on the market

Want to scale this to process 10,000 essays per day? At NexaAPI's pricing, that's a few dollars. Compare that to hiring human writing coaches.

Why NexaAPI for Ed-Tech and Writing Tools

If you're building an ed-tech tool that processes thousands of essays, cost per API call matters enormously. The difference between $0.002 and $0.020 per call is the difference between a viable product and a money pit.

ProviderGPT-4o Input (1M tokens)Free TierModels
NexaAPICheapest available✅ Yes50+
OpenAI Direct$2.50/1M tokens❌ No~15
Anthropic Direct$3.00/1M tokens❌ No~8

NexaAPI is OpenAI-compatible — just change the base URL and API key. No code rewrite needed. You get 50+ models including GPT-4o, Claude, Gemini, and open-source alternatives through one unified interface.

The Smart Developer's Response to This Research

The arXiv paper isn't a reason to avoid LLMs in education. It's a roadmap for using them correctly. Don't build AI graders. Build AI writing coaches, content generators, and feedback assistants. Use LLMs for what they're actually good at.

And when you're ready to build at scale, use the cheapest inference API available.

🚀 Start Building with NexaAPI

Reference: arXiv:2603.23714 — "LLMs Do Not Grade Essays Like Humans" (Barbosa et al., March 24, 2026) | Source retrieved 2026-03-28