Skip to content

authentic-research-partners/sciwrite-lint

Repository files navigation

sciwrite-lint

arXiv

A linter for scientific manuscripts. Checks that your references exist, your metadata is accurate, your citations support the claims you make, and your cited papers' own bibliographies are real. Works on LaTeX and PDF. Runs entirely on your machine. Produces a SciLint Score.

The only open-source tool that combines reference verification, claim checking, manuscript consistency, figure analysis, and bibliography verification in one pipeline — powered entirely by a local LLM on your own GPU.

Why

AI writing tools produce text that looks like good science — fluent prose, correct formatting, plausible-sounding citations. But they don't verify whether the references are real, whether the cited papers actually say what you claim, or whether the numbers in your abstract match your results.

sciwrite-lint does. It checks your references against academic databases, downloads the cited papers, verifies that they actually say what you claim, and follows one level deeper to check your references' own bibliographies. Fully local — no manuscripts leave your machine.

Features

22 automated checks:

Reference verification

  • Do your references exist? — checked against CrossRef, OpenAlex, Semantic Scholar, Open Library, and Library of Congress
  • Is the metadata accurate? — title, authors, year, venue compared against canonical records
  • Are any retracted? — every DOI cross-referenced against 60,000+ entries in the Retraction Watch database
  • Robust matching — when references lack DOIs, a multi-signal matching engine scores candidates across title, author, year, and venue (handles the metadata errors that LLMs routinely introduce)

Claim verification (local LLM)

  • Do cited papers support your claims? — downloads full text from 8 open-access sources (arXiv, Semantic Scholar, OpenAlex, PubMed Central, Europe PMC, Unpaywall, bioRxiv/medRxiv, CORE), parses via GROBID, embeds sections, and verifies each claim against the actual source text
  • What role does each citation play? — classifies citation purpose (evidence, contrast, method, attribution, context…) with graduated weights: an unsupported evidence citation is serious; an unsupported context citation barely matters
  • Are your references' own bibliographies real? — batch-checks cited papers' reference lists for existence, metadata accuracy, and retraction. Papers built on fabricated evidence are flagged

Manuscript consistency (local LLM)

  • Cross-section contradictions — numbers, claims, and framing that drift between sections
  • Numbers vs. tables — text claims that contradict the corresponding table or figure
  • Arithmetic and percentages — stated totals that don't match components; percentages that don't sum to 100%
  • Sample size tracking — N values that change across sections without explanation
  • Causal language — unhedged causal claims in correlational studies
  • Abstract–body alignment — abstract makes factual claims the body contradicts
  • Statistical reporting — p-values vs. their verbal interpretation
  • Structure promises — contributions promised in the introduction but never delivered

Figure checks (vision model + LLM)

  • Caption vs. content — does the caption match what the figure actually shows?
  • Text vs. figure — does the text describe the figure accurately?
  • Axis labels — units, labels, and scales consistent with the text
  • Figure–table agreement — same data in a figure and table should agree

Text checks (deterministic, no services needed)

  • Dangling citations\cite{key} with no matching bib entry
  • Dangling cross-references\ref{X} with no matching \label{X}
  • Unreferenced figures — figures included but never referenced in the text

Per-reference reliability score

All signals — metadata, retraction status, claim support, consistency, bibliography health — aggregate into a single reliability score per reference. When multiple independent checks flag the same reference, it is flagged as unreliable with specific reasons.

SciLint Score

SciLint Score = Internal Consistency × Referencing Quality × Contribution

A single number combining:

  • Internal Consistency — fraction of checks passed within the manuscript
  • Referencing Quality — are references real, accurate, and do they support your claims? Each reference weighted by its reliability score and citation purpose
  • Contribution (experimental) — five axes from philosophy of science (Popper, Lakatos, Kitcher, Laudan, Mayo): empirical content, progressiveness, unification, problem-solving effectiveness, test severity. Defaults to 1.0 until contributions runs
sciwrite-lint contributions paper.pdf --format json

Privacy and security

  • Your manuscript never leaves your machine. All parsing, LLM inference, and figure analysis run locally
  • Only citation metadata is sent externally — DOIs, titles, author names for API verification. No paper content
  • Open-weights models pinned to specific versions — results are reproducible forever, not dependent on a cloud provider's API updates
  • No API keys required — all verification uses free public databases. Optional keys increase rate limits

Two audiences

  • Humans — colored terminal output with severity levels, locations, and explanations. Decide in seconds whether each finding is real
  • AI writing agents--format json output with structured fields (level, rule_id, message, context). Run sciwrite-lint in a write → check → fix → recheck loop. Configurable exit codes for CI integration

Optimizations

Three models — a vision model (Qwen3-VL-2B default, or 8B FP8 via --vision-backend vllm for +15% accuracy), an embedding model (Arctic Embed), and an 8B reasoning LLM (Qwen3 via vLLM) — share a single consumer GPU. Models run in separate pipeline stages; on WSL2, CUDA memory virtualization pages idle allocations to system RAM, letting all three share physical VRAM. FP8 weights and KV cache (Ada Lovelace+) and per-paper SQLite caching with hash-based invalidation are baseline. On top of that:

  • Semantic section filtering — embedding-based KNN retrieval sends only the ~5 most relevant sections per claim to the LLM, reducing LLM calls ~4x
  • Prefix-first prompt structure — shared context placed before variable content in all prompts, maximizing vLLM's automatic prefix caching
  • Per-call-site thinking budgets — each LLM call site has empirically tuned (max_tokens, thinking_preset) pairs, measured via grid search to maximize detection quality
  • Adaptive embedding batches — token-aware batch sizing gives ~50x speedup over CPU while staying within VRAM limits
  • Batch-staged multi-paper pipeline — when checking 2+ papers, GPU models load once per batch (vision/embedding/cited-vision) and vLLM/network stages run concurrently, giving a meaningful speedup over sequential per-paper runs. Tune via --concurrency (default 2, validated up to 4 on a single consumer GPU)
  • Phased API resolution — citations flow through OpenAlex → Semantic Scholar → CrossRef → Open Library/LoC, each phase only processing what previous phases didn't resolve
  • 8-source full-text cascade — early exit on first successful download across arXiv, Semantic Scholar, OpenAlex, PubMed Central, Europe PMC, Unpaywall, bioRxiv, CORE
  • Live monitoring (advanced)sciwrite-lint containers monitor shows service health, VRAM usage, and KV cache utilization in a terminal dashboard

Full pipeline on a 50-reference paper: ~30 minutes initial (dominated by downloads and claim verification), minutes on cached reruns. On native Linux, the pipeline automatically swaps vLLM containers to free GPU for embedding and vision stages (~50x faster than CPU). (Native Linux GPU swap is preliminary — tested on WSL2 only; expected to work, may need minor fixes.)

Install

Assumed setup: A workstation with an NVIDIA GPU (16+ GB VRAM). Developed and tested on Windows (WSL2). Native Linux is likely to work with GPU memory allocation tuning (see docs/services.md). Not tested on macOS.

Requires uv, a container runtime (podman or docker), CUDA drivers, and NVIDIA Container Toolkit.

uv tool install sciwrite-lint --python 3.13

uv downloads Python 3.13 automatically (does not affect any Python you may already have) and installs sciwrite-lint as a globally available command.

Reproducible environment (optional)

pyproject.toml declares lower bounds for dependencies, so uv tool install / pip install may resolve newer versions as upstream releases ship. If you want the exact versions the maintainer develops and tests against — for reproducing benchmarks, debugging version-drift issues, or CI — install from requirements-pinned.txt instead.

What's in it: top-level packages declared in pyproject.toml pinned to == versions. Transitive dependencies are not pinned — pip resolves them within each top-level package's own constraints. This is a deliberate middle ground: it locks the packages sciwrite-lint actually imports and tests against, without over-constraining the dep tree.

# Clone the public repo (required — uv tool install can't read a local file)
git clone https://github.com/authentic-research-partners/sciwrite-lint.git
cd sciwrite-lint

# With uv (recommended — fast, isolated, manages Python itself)
uv venv --python 3.13 .venv
source .venv/bin/activate
uv pip install -r requirements-pinned.txt
uv pip install -e . --no-deps    # install sciwrite-lint source; deps already pinned

# Or with plain pip
python3.13 -m venv .venv
source .venv/bin/activate
pip install -r requirements-pinned.txt
pip install -e . --no-deps

Activate the venv in each new terminal with source .venv/bin/activate. For most users, the plain uv tool install sciwrite-lint above is simpler — use the pinned file only when you need exact version reproduction.

Setup

sciwrite-lint init                     # scaffold .sciwrite-lint.toml + references/ + local_pdfs/
sciwrite-lint config set-email you@example.com  # required for Unpaywall + Retraction Watch
sciwrite-lint containers start         # start GROBID + vLLM (needs GPU for vLLM)
sciwrite-lint containers monitor       # live dashboard: service health, VRAM, KV cache

Monitor dashboard

init detects .tex files and their .bib references and generates a .sciwrite-lint.toml config. Review to confirm the right files were detected.

Providing PDFs manually (local_pdfs/): Drop any reference PDF into the local_pdfs/ directory to skip the download step entirely. This is useful when you already have the file, when the paper is paywalled, or when the publisher blocks automated downloads (Cell, Springer, some PMC pages). Name files by the reference title:

local_pdfs/
├── States of Curiosity Modulate Hippocampus-Dependent Learning.pdf
├── Mind in Society.pdf
└── The System of Professions.pdf

The tool fuzzy-matches filenames against your .bib titles (threshold 0.80) and copies matches to the workspace. On the next run, matched references upgrade from T2 to T1 and go through GROBID parsing and claim verification like any other full-text reference.

Finding what to download: sciwrite-lint verify lists T2 references that are confirmed open access with direct URLs — open in your browser and save to local_pdfs/. For paywalled papers, use your institution's library access.

Optional API keys increase rate limits for Semantic Scholar, NCBI, and CORE:

sciwrite-lint config show              # see what's configured
sciwrite-lint config set-key semantic-scholar YOUR_KEY  # dedicated rate limit

See docs/services.md for GPU requirements, all external APIs, and API key details.

Usage

sciwrite-lint check --paper my_paper            # full pipeline + SciLint Score
sciwrite-lint check --paper my_paper --fresh    # same, ignoring all caches
sciwrite-lint check                             # all papers (batch-staged when 2+, ~4-5x faster than sequential)
sciwrite-lint check --concurrency 4             # batch parallelism (default 2, validated up to 4)
sciwrite-lint check paper.tex                   # text + LLM rules on a .tex file
sciwrite-lint check paper.pdf                   # check a PDF (GROBID required)
sciwrite-lint contributions --paper my_paper    # add contribution axes to SciLint Score
sciwrite-lint contributions paper.pdf           # standalone file scoring

check runs the full pipeline in one command: text checks → figure analysis → LLM consistency → reference verification via APIs → download and parse cited papers → claim verification → consistency checks on cited papers → bibliography verification → aggregate reliability scores → SciLint Score. An initial run on a 50-reference paper takes up to 30 minutes (dominated by downloads and claim verification); subsequent cached runs complete in minutes.

Use --fresh to start from scratch (backs up the existing workspace before recreating it).

Contribution axes (sciwrite-lint contributions)

contributions computes 5 contribution axes from philosophy of science (Popper, Lakatos, Kitcher, Laudan, Mayo) and updates the SciLint Score. Requires vLLM.

sciwrite-lint check --paper my_paper           # SciLint Score (contribution = 1.0)
sciwrite-lint contributions --paper my_paper   # add 5 contribution axes, update score

Individual stages

For debugging or advanced workflows, each pipeline stage is also available as a standalone command:

Command What it does
verify --paper NAME API verification only (CrossRef, OpenAlex, Semantic Scholar, Open Library, Library of Congress)
fetch --paper NAME Download full-text PDFs for verified references
parse --paper NAME Parse PDFs via GROBID and compute embeddings
verify-claims --paper NAME LLM reads cited sources, checks claim support
ref-health --paper NAME Fast reference health check: cite/bib mismatches, ID coverage, local PDF matches (no API calls)
contributions --paper NAME Add 5 contribution axes to SciLint Score (requires vLLM)

Output

Each finding has a severity level, a rule ID, and a message explaining the issue:

  • error — a concrete manuscript problem (hallucinated reference, unsupported claim, retracted source)
  • warning — needs human judgment (metadata mismatch, weak citation purpose, cross-section inconsistency)
  • info — the tool could not complete a check (LLM error, API timeout, internal crash) or informational note

Findings also carry a context field with the reasoning behind the verdict: the LLM's explanation, which identifiers were searched, or which API provided the canonical data. This lets you distinguish "the tool found a problem" from "the tool couldn't check this."

Example terminal output:

 ERROR  reference-exists        johnson2024: Not found in any API
                                Searched with: title="Deep Learning for Climate", author="Johnson"

 ERROR  claim-support           smith2023: Cited paper does not support this claim
                                Claim: "transformers outperform RNNs by 15% on BLEU"
                                Verdict: paper reports 8% improvement, not 15%

 WARN   reference-accuracy      lee2022: Year mismatch (bib: 2022, canonical: 2021)
                                Source: OpenAlex (DOI: 10.1234/example)

 WARN   cross-section-consistency
                                Abstract claims "three novel contributions" but
                                Section 5 delivers two

 WARN   reference-unreliable    chen2019: Low reliability (0.35)
                                Metadata mismatch, 2 unsupported claims,
                                23% hallucinated bibliography entries

 INFO   caption-vs-content      Figure 3: could not extract figure from PDF

  SciLint Score: 0.41
    Internal Consistency:  0.85
    Referencing Quality:   0.48
  (Run 'sciwrite-lint contributions' to add contribution axes)

Output formats: terminal (default) or --format json.

Documentation

  • sciwrite-lint checks — list all checks
  • sciwrite-lint <command> --help — detailed usage for any command
  • docs/services.md — GROBID, vLLM, external APIs, configuration

For contributors and advanced users:

Citation

If you use sciwrite-lint in your research, please cite the accompanying paper (arXiv:2604.08501):

@article{samsonau2026sciwritelint,
  title   = {sciwrite-lint: Verification Infrastructure for the Age of Science Vibe-Writing},
  author  = {Samsonau, Sergey V.},
  journal = {arXiv preprint arXiv:2604.08501},
  year    = {2026},
}

GitHub's "Cite this repository" button (powered by CITATION.cff) will also generate this for you.

License

MIT

About

A linter for scientific manuscripts

Resources

License

Stars

Watchers

Forks

Packages

 
 
 

Languages