Unjournal

Unlocking the broken black box of academic publishing with AI precision and Time-Locked transparency.

The Problem

Academic publishing is a $28 billion industry built on free labor, broken incentives, and zero accountability.

$28B Annual revenue controlled by five corporations with ~40% profit margins
10,000+ Papers retracted in 2023 alone — a record driven by systemic review failure
$0 Compensation for peer reviewers who provide the single most critical quality check
3–6 mo Average time for a single round of review — multiple rounds can take over a year

A Monopoly Built on Free Labor

Researchers write the papers, review each other's work, and serve on editorial boards — all without compensation. Publishers contribute typesetting and branding, yet extract billions. As Deutsche Bank noted: "If the process really were as complex, costly and value-added as the publishers protest, 40% margins wouldn't be available."

The Reviewer Lottery

Reviewer agreement on the same manuscript is only slightly better than chance. A paper accepted by one set of reviewers may be rejected by another. Outcomes depend on who reviews the work, not what the work contains.

A System That Silently Fails

In 2020, The Lancet and The New England Journal of Medicine published studies based on fabricated data that prompted the WHO to halt clinical trials. Peer review at two of the world's most elite journals failed to catch what any careful reader would have noticed.

How It Works

The Unjournal Pipeline: five steps that replace the broken editorial machinery.

1

Automated Quality Baseline

AI Only

Before any human sees a manuscript, AI agents check language quality, structural integrity, and formatting compliance. Only clean papers proceed to review — no more wasting expert time on typos and missing references.

2

Reviewer Identification & Targeted Outreach

AI Only

AI analyzes the manuscript's knowledge domains and searches the global publication record to find the most qualified reviewers. Each receives a personalized invitation with only the sections matching their expertise and focused questions at the intersection of the paper and their work.

3

Feedback Synthesis & Iterative Improvement

Human + AI

Human reviewers provide expert judgment. AI aggregates and synthesizes feedback into structured, prioritized reports and facilitates iterative dialogue until all substantive concerns are addressed.

4

Publication with Time-Locked Transparency

Cryptographic

The validated paper is published alongside a Time-Lock Puzzle encapsulating the complete peer review record. Confidential today, auditable tomorrow.

5

Peer Review Audit

Open to Anyone

At any point in the future, anyone can audit the peer review behind a published paper by solving its Time-Lock Puzzle. No permission needed. No institution can block it. A permanent deterrent against misconduct.

Time-Lock Puzzles

The arbiter that editors were supposed to be — but never were.

The Transparency Paradox

Open peer review sounds good in theory, but real-time transparency destroys the independence that makes expert review valuable. When reviews are visible immediately, subsequent reviewers anchor on earlier opinions. Groupthink takes hold. The signal-to-noise ratio collapses.

Time-Lock Puzzles, first formalized by Rivest, Shamir & Wagner (1996), resolve this paradox elegantly. A TLP encrypts information so it cannot be decrypted until a predetermined amount of sequential computation is performed — throwing more computers at it doesn't help.

🔒

Confidential During Review

Prevents information cascades, protects reviewer independence, and maintains the integrity of expert evaluation.

🔓

Auditable After Publication

Biased reviews, conflicts of interest, or editorial misconduct can always be discovered — by anyone, without permission from any institution.

The End of Journals

Unjournal doesn't compete with journals. It makes them unnecessary.

No Rejections

Every submission enters a structured improvement process. Authors work with relevant experts until their research meets community standards. The journal becomes a mentor, not a filter.

No Journal Silos

Peer review becomes a universal process, not a per-journal process. A review conducted for one submission is never wasted. Expert labor is respected and preserved.

No Prestige Gatekeeping

The quality of a paper is determined by the rigor of its review, not by the brand of the journal that accepted it. Science speaks for itself.

Business Model

Transparent fees. Aligned incentives. Fair compensation.

Unjournal operates as a SaaS platform. Authors and institutions pay a transparent fee for review orchestration — replacing the opaque, inflated article processing charges (APCs of $2,000–$11,000) extracted by legacy publishers.

The model aligns incentives correctly: Unjournal succeeds when authors get better reviews faster, not when more papers are rejected or more subscriptions are sold.

Reviewer compensation is built into the service fee. By making the cost of peer review explicit, Unjournal opens the door to fairly compensating reviewers for their expertise — a possibility the current system structurally prevents.

Stakeholder Legacy System Unjournal
Authors Pay $2K–$11K APCs Transparent, lower fees
Reviewers $0 compensation Fair pay for expertise
Quality 2–3 generalist reviewers 10+ targeted specialists
Speed 3–12 months Days to weeks
Accountability Black box, unauditable Cryptographic auditability

Science deserves a review process as rigorous as the research it evaluates.

Unjournal is building the future of peer review. Join us.

We'll notify you when early access opens. No spam, ever.