Benchmarking OCR Accuracy on Scanned Contracts, Invoices, and Forms
A practical OCR benchmarking framework for contracts, invoices, and forms across scan quality and preprocessing settings.
Benchmarking OCR Accuracy on Scanned Contracts, Invoices, and Forms
OCR benchmarking is only useful when it reflects how documents fail in the real world. A lab-perfect scan of a clean invoice tells you little about how your pipeline will handle a skewed contract, a faxed form, or a thermal receipt with fading text. In practice, teams need a repeatable way to measure accuracy across document types, scan quality levels, and preprocessing settings so they can compare engines fairly and choose the right workflow. If you are building or evaluating a production pipeline, this guide pairs benchmark design with deployment realities such as secure OCR ingestion workflows, regulatory change management, and audit-friendly logging.
We will define a benchmark framework, specify the metrics that matter, and show how to interpret results for invoice OCR, contract OCR, and form OCR under different preprocessing strategies. We will also cover how to account for document variability, why precision and recall can move in opposite directions, and how to avoid misleading “overall accuracy” claims. For teams planning an enterprise rollout, this kind of benchmarking belongs alongside your integration plan, just as you would benchmark any other data pipeline before production.
1. What OCR Benchmarking Should Measure
Document-level accuracy versus field-level accuracy
OCR accuracy is not a single number unless your documents are trivially simple. For scanned contracts, invoices, and forms, you need to distinguish document-level correctness from field-level extraction quality. Document-level accuracy asks whether the output is usable end to end, while field-level accuracy evaluates specific items such as vendor name, invoice total, signature block text, clause references, or form checkbox state. In most enterprise scenarios, field-level accuracy is more actionable because one missing invoice total can be more expensive than ten minor punctuation errors.
A useful benchmark separates the document into logical units. For invoices, that usually means header, line items, totals, tax fields, and remittance details. For contracts, it means parties, effective date, term, obligations, and signature areas. For forms, it means labels, values, tables, and whether a field was handwritten or machine printed. If you also need guidance on downstream extraction design, see our high-frequency dashboard design and team collaboration patterns for AI operations.
Precision, recall, and character error rate
For OCR, precision and recall often matter more than a simplistic “percent correct.” Precision tells you how many extracted tokens or fields were correct out of all extracted items; recall tells you how many ground-truth items were recovered. High precision with low recall usually means the engine is conservative and misses content, which is dangerous for legal or finance workflows that require completeness. High recall with low precision means you get too much noise, which can create hidden cleanup costs and false automation.
Character error rate, word error rate, and field accuracy should be used together. Character error rate helps compare raw OCR quality at the text layer, while word error rate is better for readable text and line-oriented content. Field-level exact match is essential for structured extraction tasks such as invoice OCR and form OCR. When benchmark results are tied to operational context, they help teams understand whether a slight improvement in CER actually reduces manual review time or whether it is just statistically interesting. For related security-oriented evaluation methods, review incident runbook design and data protection against AI misuse.
Why “accuracy” alone is misleading
Two OCR systems can both claim 98% accuracy and still behave very differently on real documents. One might be strong on clean machine-printed invoices but collapse on multi-column contracts, while another may handle poor scans well but struggle with precise table extraction. Accuracy also depends on the tolerance rules you define: do you count case differences, punctuation, whitespace normalization, or merged line breaks as errors? Benchmarks that do not spell out these rules usually produce results that cannot be reproduced or trusted.
That is why a benchmark framework should always define token normalization, field matching rules, and scoring thresholds before the first sample is processed. If you are reporting to procurement or compliance stakeholders, that rigor is critical. It mirrors the discipline used in regulated workflows such as HIPAA-conscious ingestion and in review-heavy procurement systems like public-sector solicitation processes, where documentation quality determines whether an item moves forward.
2. Building a Benchmark Dataset That Actually Reflects Production
Balance document types, layouts, and content density
A credible benchmark dataset should include representative samples of contracts, invoices, and forms from your actual operating environment. Contracts often contain dense prose, clause numbering, signatures, and varying typography. Invoices usually mix free text with tables, totals, and IDs. Forms combine fixed fields, handwritten values, checkboxes, and occasional stamps or annotations. If your benchmark only includes one clean template per type, your results will not survive contact with production.
To make comparisons fair, build a balanced dataset with enough variety to stress the OCR engine. Include multi-page contracts with scans from different devices, invoices with and without tables, and forms that vary by handwriting quality and field density. If you need inspiration for thinking about operational variability, consider how market research and competitive intelligence rely on representative sampling before drawing conclusions. The same logic applies here: the sample has to resemble reality, not ideal conditions.
Define ground truth carefully
Ground truth is the foundation of every OCR benchmark. For each sample, manually verify the exact text, field values, and structural annotations you expect the OCR system to recover. This is not just about typing the text into a spreadsheet; it includes marking reading order, table boundaries, checkbox state, and whether a string is intentionally ambiguous. In contract OCR, for example, you may need to decide whether section headers and footer disclaimers count as extractable text or should be ignored.
Use double review for high-stakes samples. One reviewer can transcribe, another can validate, and disagreements can be resolved with an adjudication pass. This matters because benchmark noise inflates or deflates accuracy in ways that look like product changes but are actually dataset defects. Teams that handle sensitive records often apply similar discipline when preparing archives for search, e-discovery, or analytics, especially in workflows tied to medical record ingestion and device logging.
Annotate scan quality levels explicitly
Instead of mixing all scans into one bucket, segment them by quality level. A practical scale includes clean, normal, degraded, and severe. Clean might mean straight pages, high contrast, and 300 DPI or better. Normal could include mild skew or light background noise. Degraded might feature faint text, compression artifacts, low contrast, or slight blur. Severe can include cropped edges, heavy skew, shadows, and low-resolution fax-like output. This classification lets you answer an operational question: at what quality level does preprocessing stop helping and manual review become cheaper?
Annotating scan quality also prevents benchmarking from becoming a “mean of mismatched inputs.” A model that excels on clean scans but degrades sharply on severe scans may still be the best option if your capture environment is controlled. Conversely, a slightly weaker engine on clean pages might outperform overall if it is far more robust to noisy scans. This kind of real-world tradeoff is similar to what teams weigh in security device evaluation and tool selection, where edge-case performance determines practical value.
3. Document-Specific Benchmark Design
Invoice OCR: totals, tables, and line items
Invoice OCR should be scored by business-critical fields, not by how pretty the extracted text looks. At minimum, benchmark vendor name, invoice number, invoice date, subtotal, tax, total, currency, purchase order number, and line-item descriptions. If your workflow uses invoice automation, line-item precision and recall often determine whether your accounts payable team trusts the output. A system that gets totals right but drops a line item can still create reconciliation problems, duplicate payments, or delayed approvals.
For invoices with tables, benchmark both text extraction and table reconstruction. Line breaks, merged cells, and column order can change the meaning of the invoice. You should test whether preprocessing improves table structure or accidentally damages it. In many cases, aggressive thresholding helps character recognition but hurts table geometry. To understand how structured data quality affects business decisions, compare your results with operational frameworks like cost model analysis where a few bad inputs can distort the full downstream calculation.
Contract OCR: long-form text, clauses, and signatures
Contract OCR is usually harder than invoice OCR because contracts are longer, denser, and more layout-diverse. Important benchmark targets include parties, dates, governing law, renewal clauses, notice language, exhibit references, and signature blocks. You should also score whether the engine preserves reading order across columns, footnotes, and page headers. If a model loses a clause heading or merges adjacent sections, downstream legal search and clause analytics can fail even if the text itself is partially correct.
Contracts are especially sensitive to preprocessing because legibility and structure compete. Dewarping can recover line order on curved pages, but overcorrection may distort small-print annotations or stamp overlays. Deskewing helps most scans, but too much rotation can harm page segmentation. If you are evaluating contract pipelines in a regulated or procurement-heavy environment, also review how document amendments and approvals behave in workflows like the VA solicitation process, where incomplete signatures or missing revisions can block progress. That operational parallel is a useful reminder: contract OCR must preserve both content and compliance cues.
Form OCR: fields, checkboxes, and handwriting
Form OCR is the most structured of the three, but it can be deceptively difficult because field boundaries are often fixed while content quality varies wildly. Benchmark the extraction of labels, values, checkboxes, radio buttons, dates, and signatures. If handwriting is present, split results between printed and handwritten fields, because the model may be strong at one and weak at the other. Many production errors come from adjacent fields bleeding into each other rather than from pure character mistakes.
For forms, the benchmark should include whether the OCR engine respects field association. Extracting “John Smith” is not enough if it is attached to the wrong field label. You should also measure false positives on empty fields, because hallucinated values can be more harmful than missing ones. For design patterns related to structuring form data, see designing identity dashboards and collaboration workflows for AI teams, both of which emphasize clarity, traceability, and clean state transitions.
4. Scan Quality Levels and How to Test Them
Resolution, contrast, skew, and noise
Scan quality is the easiest place to introduce hidden bias into OCR benchmarks. Resolution affects character shape; contrast affects foreground-background separation; skew affects segmentation; and noise affects both character detection and language-model postprocessing. A 300 DPI scan of a clean contract will behave very differently from a 150 DPI fax of the same page. Benchmarks should therefore vary each parameter independently when possible, rather than only comparing “good scan” versus “bad scan.”
Use a grid of quality settings so you can attribute gains or losses accurately. For example, test 200, 300, and 400 DPI; low, medium, and high compression; and mild, moderate, and severe skew. This reveals the point at which preprocessing helps enough to justify its cost. It also prevents teams from overfitting their recommendations to a narrow capture device or scanner brand. In procurement terms, this is the difference between an anecdotal win and a defensible evaluation.
Real-world degradation patterns
Not all bad scans fail in the same way. Contracts often have shadows from binding, page curls, or marginal notes. Invoices may have thermal fading, dot-matrix artifacts, or stamps that obscure totals. Forms often suffer from uneven handwriting pressure, field crowding, and photocopy blur. Your benchmark should include these degradation patterns because they trigger different OCR failure modes and require different remediation tactics.
One useful method is to label each sample with one primary defect and one secondary defect. That makes it possible to isolate why performance changed after preprocessing. For example, a thresholding step may help faded invoices but not improve low-contrast contracts. This is similar to how teams analyze risk factors in regulatory change management or security incident planning: the issue is rarely one-dimensional, and the response has to fit the failure mode.
Benchmarking with capture-device diversity
If your intake comes from multiple scanners, MFPs, mobile cameras, and uploaded PDFs, include each source in the benchmark. OCR engines often perform differently on camera-captured images because of perspective distortion, autofocus artifacts, and variable lighting. Mobile capture also introduces cropping inconsistencies that scanner-based pipelines do not see. A truly useful benchmark tells you how much of your accuracy comes from the OCR engine versus the capture path.
This source diversity matters for migration planning too. A team moving from a paper-first workflow to a distributed scan-and-upload process should not assume the same OCR settings will work across all input channels. Good benchmark data gives you the evidence to choose different preprocessing profiles by source, much like how competitive benchmarking separates segments before making strategic decisions.
5. Preprocessing Settings: What to Test and Why
Deskewing, dewarping, and denoising
Preprocessing can materially improve OCR accuracy, but only if you measure its effect systematically. Deskewing is usually beneficial for all document types because it stabilizes line detection and reduces segmentation errors. Dewarping is especially helpful for book-bound contracts or curled scans, but it may introduce artifacts if applied to already-flat pages. Denoising can improve character recognition on noisy scans, though excessive smoothing may blur punctuation or thin fonts.
Benchmark each preprocessing step independently and in combination. For example, compare raw input, raw plus deskew, raw plus denoise, and raw plus deskew plus denoise. Then test whether the gains vary by document type and scan quality. In many real deployments, the best outcome is not the most aggressive preprocessing, but the smallest set of transformations that consistently lifts recall without sacrificing precision. This is especially relevant in workflows where speed matters and every added step increases latency.
Thresholding, binarization, and contrast normalization
Thresholding can dramatically improve OCR on faded documents, but it is also one of the easiest ways to destroy subtle detail. Global thresholding may work on uniform invoices but fail on documents with mixed background shading. Adaptive thresholding is more resilient, especially on scanned forms and contracts that contain stamps, logos, or uneven lighting. Contrast normalization can help text stand out, but it must be tuned carefully to avoid oversaturation.
When comparing thresholding settings, use both text-level and field-level metrics. A processing step that lowers character error rate may still hurt checkbox detection or table border preservation. This is why preprocessing benchmarks should report per-document-type deltas, not just aggregate averages. If you need a security-aware perspective on handling sensitive scanned data during preprocessing, see cloud data protection guidance and logging controls for business systems.
When preprocessing hurts more than it helps
It is common for teams to assume that more preprocessing equals better OCR. In reality, a well-trained engine may already handle mild noise, and extra processing can degrade geometry or remove useful texture. Overprocessing often shows up in form OCR as broken field boxes, in contract OCR as lost superscript or footnote markers, and in invoice OCR as merged digits. If your benchmark does not include a no-preprocessing baseline, you may never realize the engine was already near optimal on some inputs.
Pro tip: treat preprocessing as an ablation study, not a magic fix.
Pro Tip: Start with the smallest preprocessing set that improves recall on your noisiest 20% of documents. If a setting helps only clean scans, it is probably not the right default for production.
6. A Practical Benchmark Framework You Can Reuse
Step 1: Define scope and success criteria
Begin by selecting the document classes, field sets, and business thresholds you care about. For invoice OCR, that might mean 99% recall on invoice totals and 95% exact match on vendor name. For contracts, success might be clause-level recall above a threshold for search indexing. For forms, it may be zero false positives on blank required fields. A benchmark without explicit success criteria becomes a vanity exercise, because no one can tell whether a difference matters operationally.
Write the benchmark scope down before you run anything. Include the document types, scan quality bands, preprocessing configurations, and evaluation metrics. Also define the acceptable tradeoff between accuracy and speed, because many teams do not need the absolute best OCR if the latency or cost is too high. This is where benchmarking connects to product evaluation and pricing research, much like the logic behind product and pricing research.
Step 2: Run controlled comparisons
Execute the same dataset against each OCR configuration under identical scoring rules. If you are comparing engines, keep preprocessing constant. If you are comparing preprocessing, keep the engine constant. If you vary both at once, you will not know which change caused the gain. Controlled comparisons are the only way to produce credible results that developers, IT managers, and procurement teams can trust.
Use confidence intervals or at least repeated runs when stochastic components are involved. Some OCR pipelines include layout analysis, language model postprocessing, or adaptive heuristics that can vary slightly. If your variation is due to randomization, then single-pass results may overstate or understate true performance. In regulated or security-sensitive environments, that kind of uncertainty is a liability, so measure carefully and document the test conditions.
Step 3: Segment and analyze error modes
After scoring, inspect errors by document type, quality level, and defect type. Group by missing text, wrong text, field misassignment, table structure failure, and reading-order issues. This tells you whether the engine’s weaknesses are concentrated in one part of the workflow or spread across everything. A strong overall score with one catastrophic failure mode is usually not production-safe.
For example, if the engine has high recall on contracts but repeatedly misreads signature blocks, you may need a targeted preprocessing or postprocessing rule. If invoices are mostly correct except for line-item splitting, you may need table-aware extraction. If forms show many false positives on empty fields, you may need stricter confidence thresholds. Similar segmentation is common in medical document ingestion and security operations, where failure modes are mapped before remediation.
7. Interpreting Results for Production Decisions
Choose the right metric for the workflow
There is no universally “best” OCR metric. If your workflow is search and discovery, recall may matter most because missing text is worse than a few false extractions. If your workflow is auto-approval or payment posting, precision matters more because false positives can trigger costly mistakes. For human-in-the-loop review systems, the right metric may be review load reduction, not raw OCR score, because the goal is to minimize the amount of text that requires manual correction.
Connect the metric to the business process. Invoices often need high field precision, contracts need high semantic recall, and forms need exact field mapping. If you are still unsure, compare against downstream error costs rather than generic accuracy scores. That framing is much more useful to developers and IT administrators who need to justify deployment choices internally.
Understand the cost of false positives and false negatives
A false negative on a contract clause might mean a search miss, an incomplete compliance review, or a failed redline comparison. A false positive on an invoice total might mean an incorrect payment or accounting exception. A false positive on a form field can trigger downstream workflow automation with garbage data. The financial and operational cost of each error type should shape your threshold settings and your decision to automate or route to review.
One practical tactic is to set separate thresholds by document section. For example, allow lower confidence on invoice descriptions but require higher confidence on totals. On contracts, allow more aggressive capture of body text but stricter rules for signature dates. On forms, require near-perfect accuracy for identifiers and more permissive rules for optional fields. This kind of tiered logic often beats a one-size-fits-all threshold.
Benchmark latency and throughput alongside accuracy
Accuracy alone can mislead teams into choosing an OCR workflow that is too slow for real operations. If preprocessing doubles accuracy but triples latency, it may still be a bad tradeoff for high-volume invoice ingestion. Benchmark throughput, time per page, and end-to-end processing time alongside OCR quality. Production success usually depends on the balance between correctness and operational cost.
This is especially important when you are evaluating SDKs, hosted APIs, or hybrid deployments. A system that is slightly less accurate but significantly faster can outperform in practice if the workflow includes manual review. As with other enterprise systems, such as the careful evaluations described in health-tech security checklists and security submission guidance, the best choice is the one that fits your operational constraints.
8. Example Benchmark Table: How to Compare OCR Configurations
The table below shows a practical way to compare OCR configurations across document type, scan quality, and preprocessing. Use it as a template, then replace the illustrative scores with your own measurements. The key is to keep the structure consistent so each result has the same context and can be reproduced later.
| Document Type | Scan Quality | Preprocessing | Primary Metric | Illustrative Outcome |
|---|---|---|---|---|
| Invoice | Clean 300 DPI | None | Field F1 | Very high totals accuracy; line items mostly correct |
| Invoice | Degraded 200 DPI | Deskew + adaptive threshold | Field F1 | Totals improve, line-item recall rises |
| Contract | Normal 300 DPI | Deskew only | Clause recall | Reading order preserved; legal clause capture stable |
| Contract | Severe skew and shadow | Dewarp + denoise | Word error rate | Body text improves, signature block still needs review |
| Form | Mixed print/handwriting | Contrast normalize | Field exact match | Printed fields strong; handwritten fields remain variable |
| Form | Low contrast photocopy | Deskew + binarization | Checkbox precision | Checkbox detection improves, blank-field false positives drop |
Use this table style in your own internal evaluations so stakeholders can compare apples to apples. If your team manages data pipelines across multiple business units, standardizing the reporting format is just as important as the OCR results themselves. It reduces debate and makes it easier to pick a path forward.
9. Common Benchmarking Mistakes to Avoid
Using one sample set for all workflows
Do not use a single generic sample set to evaluate invoices, contracts, and forms as though they were interchangeable. Each document type has different layout rules, failure patterns, and business-critical fields. When you mix them without segmentation, you hide the engine’s weaknesses and create false confidence. A model that performs well on one type can still be a poor choice for another.
Ignoring postprocessing and human review
OCR output is often just the first stage in a larger workflow. If your benchmark ignores normalization, validation, field mapping, and human review, the numbers may not reflect operational reality. In practice, a pipeline with slightly lower raw OCR accuracy can outperform if its downstream cleanup is better designed. That is why benchmarking should include the full path from scan to structured output, not only the text layer.
Failing to document scoring rules
If your benchmark does not state how text is normalized, how tables are scored, and how empty fields are treated, others cannot reproduce your results. Reproducibility is essential when you are comparing vendors or justifying deployment choices to leadership. Good documentation also makes future re-benchmarking easier when you change scanners, update preprocessing, or switch OCR models. In practice, this is what separates a one-off test from a durable evaluation framework.
10. Practical Recommendations for Developers and IT Teams
Start with the highest-volume, highest-risk documents
Do not begin your benchmark with the easiest sample set. Start where OCR failure would hurt most: invoices with many line items, contracts with long clauses, or forms that trigger automated decisions. Those documents expose the gaps that matter to operations and finance. Once you know the weak spots, you can decide whether to tune preprocessing, add manual review, or switch engines.
Build a benchmark harness, not a spreadsheet habit
It is fine to prototype in spreadsheets, but a serious OCR evaluation should eventually live in a reproducible harness. The harness should load samples, apply preprocessing, send requests to the OCR engine, score outputs, and store results for comparison over time. This turns benchmarking into a routine regression test, which is essential when you upgrade scanners, modify templates, or change OCR models. Teams that treat benchmarking as a one-time project often discover quality drift only after production users complain.
Use benchmark results to guide deployment policy
Finally, turn benchmark findings into concrete workflow rules. You might route severe-quality invoices to manual review, allow clean contracts to auto-index, or require handwriting-specific handling for forms. That policy layer is where benchmarking pays off, because the result is not just better numbers but better decisions. For more on building secure, operationally aware document pipelines, revisit OCR ingestion workflows, regulatory readiness, and logging and monitoring controls.
11. Decision Checklist: What a Good OCR Benchmark Should Answer
Is the dataset representative?
Your benchmark should reflect real document types, scanner sources, and degradation patterns. If not, the numbers will not predict production quality. Ask whether the sample mix matches your actual workflow volume, not just your ideal case. A representative dataset is the difference between a useful benchmark and a marketing demo.
Do the metrics map to business risk?
Choose metrics that match the cost of OCR mistakes. If a missed invoice total is unacceptable, measure total-field recall and precision directly. If contract search is the goal, emphasize clause recall and reading-order preservation. If blank form fields must never be hallucinated, track false positives aggressively.
Can the benchmark be rerun after changes?
Rerun capability is essential because OCR performance changes with engine updates, preprocessing tweaks, and document mix shifts. A benchmark that cannot be rerun loses its value very quickly. Make the process repeatable, version-controlled, and well documented so it becomes part of your release process.
Conclusion
A useful OCR benchmark is not a single score; it is a framework for understanding how accuracy changes across document type, scan quality, and preprocessing choice. Contracts, invoices, and forms each fail in distinct ways, so they must be measured separately and interpreted in the context of downstream business risk. Precision, recall, field accuracy, table integrity, and reading order all matter, and the right preprocessing can help or hurt depending on the scan conditions. The goal is not to chase the highest number on a slide deck, but to identify the configuration that gives your team the best operational outcome.
If you are planning a pilot or replacing an existing OCR stack, use this framework to build a benchmark that your developers, IT team, and stakeholders can trust. Then pair it with secure, well-documented implementation practices from guides like HIPAA-conscious OCR workflows and incident response runbooks. That combination—measurement plus operational discipline—is what turns OCR from a pilot into a dependable production system.
Related Reading
- How to Build HIPAA-Conscious Medical Record Ingestion Workflows with OCR - Secure document ingestion patterns for sensitive healthcare data.
- Understanding Regulatory Changes: What It Means for Tech Companies - A practical view of compliance-aware product planning.
- Understanding the Intrusion Logging Feature - Why audit trails matter in document automation systems.
- How to Build a Cyber Crisis Communications Runbook for Security Incidents - A structured approach to operational readiness.
- Market Research & Insights - How benchmark discipline mirrors credible market evaluation.
FAQ
What is the best metric for OCR accuracy?
The best metric depends on the workflow. For free text, character error rate and word error rate are useful. For structured data, field-level precision, recall, and exact match are usually better. For invoices and forms, business-critical field accuracy is often the most important metric.
Should preprocessing always improve OCR results?
No. Preprocessing can help on noisy or skewed scans, but it can also damage layout, blur thin characters, or create artifacts. The only reliable way to know is to benchmark preprocessing options against a no-preprocessing baseline.
How many documents do I need for a meaningful benchmark?
Enough to represent your real document mix and quality range. In practice, that means multiple samples per document type, per scan quality band, and per major template or layout family. More important than raw size is coverage of the failure modes you care about.
Why do contract, invoice, and form OCR need separate benchmarks?
Because each document type has different structure and error patterns. Contracts are long and clause-heavy, invoices are table-heavy, and forms are field-structured with handwriting and checkboxes. A single blended score hides these differences and makes deployment decisions less reliable.
How do I compare OCR engines fairly?
Use the same dataset, same ground truth, same normalization rules, and same preprocessing when comparing engines. Then report metrics by document type and scan quality so you can see where each engine is strong or weak. Fair comparisons are controlled comparisons.
Related Topics
Daniel Mercer
Senior SEO Content Strategist
Senior editor and content strategist. Writing about technology, design, and the future of digital media. Follow along for deep dives into the industry's moving parts.
Up Next
More stories handpicked for you
OCR for Financial Market Intelligence Teams: Extracting Tickers, Options Data, and Research Notes
How to Build an OCR Pipeline for Market Research PDFs, Filings, and Teasers
Comparing OCR Accuracy on Medical Records: Typed Forms, Handwritten Notes, and Mixed Layouts
Data Residency and OCR for Health Records: What Teams Need to Know
Designing Consent and Access Controls for OCR in Sensitive Health Workflows
From Our Network
Trending stories across our publication group