PDF→Markdown Evaluation

9 PDFs · 900 questions · Multiple parsers (Mathpix / TextIn / Reducto / Marker / Internal)

Last updated: 2025-08-12 22:39 UTC

Datasets (9)

How this works

  1. Convert each PDF with multiple parsers → store Markdown under docs/md/<parser>/<doc_id>/
  2. 100 PDF-specific questions per doc → questions/<doc_id>/questions.jsonl (mirrored to site)
  3. Run RAG pipelines → write answers to runs/<parser>/<doc_id>/answers.jsonl
  4. Use an LLM evaluator (or exact-match) against the original PDF to score