From Patient Portal PDFs to Searchable Intelligence: A Healthcare Document Workflow
Turn patient portal PDFs into searchable, structured healthcare records with OCR, metadata, indexing, and AI summaries.
Patient portals have made it easier than ever for patients to upload forms, discharge summaries, lab reports, referral letters, and insurance documents. The problem is that most of those files arrive as PDFs or image scans, which means they are visible to humans but still invisible to systems. For healthcare teams, that creates a bottleneck: staff manually open files, find key facts, copy them into records, and then hope nothing was missed. For patient-facing apps, it creates an even bigger issue because the app may receive documents it cannot classify, search, or summarize in a reliable way.
This guide shows how to turn that messy document intake stream into structured, searchable intelligence using OCR, metadata extraction, document indexing, and AI summarization. It is designed for technology professionals, developers, and IT admins building healthcare workflows that must be fast, accurate, privacy-first, and practical to integrate. The core model is simple: ingest documents from the patient portal, extract text and fields, assign metadata, index the output, and route it into the right system of record. If you are already thinking in terms of automation, governance, and scale, this is the workflow to implement. For a broader view of how AI is reshaping enterprise document operations, see agentic AI in document workflows.
OpenAI’s recent healthcare-facing product launch also shows how quickly demand is shifting toward AI-assisted medical record review, while reminding us that privacy and data separation must remain non-negotiable. Health information is among the most sensitive classes of data, so any workflow must be designed with strict boundaries, access control, and clear retention policies. That balance between usefulness and safety is exactly why a purpose-built document workflow matters. It lets teams gain operational efficiency without depending on ad hoc human review or exposing sensitive records to unnecessary risk. The same privacy-first mindset that applies to modern AI health tools also applies to your intake pipeline, your indexing layer, and your search experience.
1. Why patient portal documents break traditional workflows
Document uploads are human-readable, not system-ready
Most patient portal uploads arrive in forms that are easy for a person to inspect but hard for software to process. A scanned referral might contain typed text, handwritten notes, stamps, and a rotated page in the same file. An insurance explanation of benefits may include multiple columns, nested tables, and small print that breaks naive parsing. Without OCR and extraction logic, your workflow becomes a chain of manual review tasks that wastes time and introduces errors.
The operational cost is not just labor. Every manual handoff creates latency, and latency can delay prior authorizations, referral triage, patient onboarding, and care coordination. It also reduces consistency because different staff members may interpret the same document differently. In a healthcare workflow, inconsistency is more than inefficient; it can affect downstream decisions and patient experience.
Searchability is a workflow requirement, not a nice-to-have
Once documents are stored in a portal or records system, teams need to answer questions quickly: What was the patient’s last A1C? Which specialist requested the referral? Is there a signed consent form? If the answer requires opening each PDF manually, your archive is storage, not intelligence. Searchable records should support full-text search, field-level filters, and document-level metadata so users can find relevant files in seconds.
That is why document indexing is central to the solution. Indexed content allows staff to search across the content of documents, not just the filename. It also unlocks use cases like duplicate detection, missing-document checks, and patient timeline reconstruction. If you want to understand how better indexing improves retrieval and knowledge workflows, it helps to think of it as a healthcare version of enterprise content discovery.
Patient-facing apps need instant answers, not file dumps
Patients increasingly expect digital experiences that do more than display a list of uploaded files. They want plain-language summaries, context-aware search, and guided next steps. A patient portal that merely stores PDFs forces the user to decode their own records, which is frustrating and often impossible for non-clinical users. The better approach is to generate structured insights from the document and make them available as labels, summaries, and links back to source pages.
This is where AI summarization can help, but only after the text has been extracted accurately. Summaries should not replace source documents; they should accelerate understanding. For best results, the system should preserve the original scan, extract the text, and then produce a controlled summary that can be reviewed or constrained by rules. That design keeps the experience useful while reducing the chance of hallucinated medical claims.
2. The healthcare document workflow, step by step
Step 1: Ingest from the patient portal
The workflow begins when a user uploads a document through a patient portal, provider app, claims system, or staff inbox. At ingestion time, capture the file, the uploader identity, the timestamp, and any form context such as visit type or department. This initial metadata matters because it helps route the file and sets the foundation for downstream indexing. A good ingestion pipeline should also normalize file types so PDFs, JPEGs, PNGs, and multi-page TIFFs can all be handled consistently.
In healthcare, ingestion is also a governance checkpoint. The system should immediately classify the upload as potentially protected health information and apply the right access controls. If your workflow includes pre-processing on edge infrastructure, you can reduce exposure by minimizing how long raw files remain in transient storage. For deployment patterns that push compute closer to the source, see edge AI for DevOps.
Step 2: OCR and PDF extraction
Once a document arrives, OCR converts the visual text into machine-readable text. For native PDFs, text extraction can sometimes happen directly without OCR, but real-world healthcare files often mix embedded text, scans, and images. A robust system should detect page type automatically and apply the right extraction method for each page. That hybrid approach improves speed and usually yields better accuracy than treating every file the same way.
Accuracy matters most on the documents that are hardest to parse: handwritten intake forms, faxed referrals, lab printouts, and low-resolution scans. If your workflow supports a privacy-first OCR API, you can integrate extraction without building your own recognition stack. That helps teams move faster while keeping architecture simple. If you are weighing operational costs and throughput tradeoffs, it is worth reviewing a practical benchmark approach like secure cloud data pipelines.
Step 3: Extract fields and normalize metadata
After raw text extraction, the next layer is structure. The system should identify high-value fields such as patient name, date of birth, document type, provider, specialty, date of service, medication names, diagnosis codes, and signature presence. Normalization means turning varied representations into standard values, such as converting all dates to ISO format and all provider names to canonical IDs. Once structured, this data can be used for routing, indexing, filtering, and QA.
Metadata is the difference between a document archive and a usable workflow. A scanned referral with no metadata is just a file. The same referral with document type, originating clinic, urgency flag, and extracted diagnosis becomes an operational object that can drive downstream actions. For teams building a broader knowledge layer from structured content, the lessons from knowledge management systems apply surprisingly well here.
Step 4: Index into searchable records
Once text and metadata are available, index them into a search engine or document store designed for retrieval. Good indexing supports both exact-match and semantic search. Exact-match search is critical for item lookups like lab dates or ID numbers, while semantic search helps staff find relevant records even when the wording varies. The index should store document-level metadata, page-level references, and optionally paragraph-level chunks for more precise retrieval.
Searchable records should always include a path back to the original document and page number. In healthcare, traceability is essential because staff need to validate extracted data against the source. If the output is used in a patient-facing app, that traceability becomes even more important. Users should be able to click a summary field and see the exact page where that fact came from.
Step 5: Summarize and route
AI summarization should sit after extraction and indexing, not before. When text is cleanly extracted and segmented, you can generate useful summaries such as “Referral from cardiology requesting follow-up within 2 weeks” or “Discharge note mentions medication change and next appointment date.” These summaries improve triage speed, reduce reading time, and make patient-facing interfaces far more usable. The summary layer can also power alerts, work queues, and automated next-step recommendations.
To avoid overclaiming, summaries should be bounded by source text and clinical rules. They should not invent diagnoses or make treatment recommendations. This is especially important given the broader caution around consumer-facing AI health tools and the need for airtight separation of sensitive information. As the BBC reported in its coverage of OpenAI’s health feature, trust depends on strong safeguards and clear boundaries around how medical records are handled.
3. Architecture patterns that actually work in healthcare
Pattern A: Batch intake for staff review
Batch intake works well for providers, clinics, and back-office operations where uploaded documents are processed every few minutes rather than in real time. In this model, a queue ingests portal uploads, OCR processes them asynchronously, and staff see the results in a review dashboard. The advantage is operational simplicity: failures can be retried, and quality control can happen before the record is finalized. This is a strong choice when accuracy and traceability matter more than instant response.
Batch intake is especially useful for records management teams and revenue cycle workflows. It lets you segment documents by priority and process high-value files first, such as authorizations or urgent referrals. It also reduces the risk of overloading downstream systems during peak upload windows. If your team is formalizing this architecture, the ideas in agentic-native SaaS operations can help frame automation without losing control.
Pattern B: Real-time patient app enrichment
Real-time enrichment is the right pattern when the portal experience needs immediate feedback. For example, after a patient uploads a discharge summary, the app may instantly display extracted document type, encounter date, and a short summary. This reduces uncertainty and gives users confidence that the upload succeeded. In more advanced implementations, the app can also suggest the next step, such as “awaiting review” or “missing signature page.”
The challenge with real-time workflows is latency management. You need fast OCR, lightweight link handling, and reliable fallback states if a file is still processing. The user experience should be transparent about status rather than forcing a generic spinner. That transparency reduces support tickets and helps the app feel trustworthy rather than opaque.
Pattern C: Hybrid clinical and administrative pipelines
Many healthcare organizations need both patient-facing convenience and internal operational control. A hybrid pipeline can send the same uploaded document into two paths: one for clinical review and one for administrative indexing. The clinical path might focus on care-relevant entities and summary snippets, while the administrative path extracts IDs, coverage details, and signature status. This dual-route approach eliminates duplicate manual work.
Hybrid routing also improves governance because each team sees only what they need. Role-based views can hide sensitive content from users who do not require it, while preserving full fidelity for authorized staff. That separation mirrors best practices in secure data handling and reduces the risk of overexposure. If your organization is modernizing its collaboration model, the cautionary lessons from AI risk management are worth studying.
4. Metadata design: the difference between retrieval and chaos
Core metadata fields to capture
At minimum, healthcare document workflows should capture document type, patient identifier, source channel, upload timestamp, OCR confidence, page count, and extraction status. For medical operations, add provider name, encounter date, department, and any detected form category such as consent, referral, imaging, or discharge. If the document is tied to claims or revenue cycle processes, capture payer, plan number, and reference IDs as well. These fields make downstream automation much easier.
It is also valuable to track page-level metadata when documents are multi-page and heterogeneous. A single PDF can contain a consent form on page one and a lab report on page two. Treating the entire file as one blob limits search precision and can make retrieval noisy. Page-level intelligence lets you route individual pages to the correct team or workflow branch.
Confidence scores and human review thresholds
Not every extracted field should be accepted automatically. Confidence scoring helps determine when the system can auto-index a value and when a reviewer needs to verify it. For example, a clearly printed patient name may meet a high-confidence threshold, while a handwritten dosage may be flagged for review. This is a practical way to balance speed and quality in real healthcare settings.
Human-in-the-loop review is not a weakness; it is a design choice. The best systems use automation to remove repetitive work while preserving review for edge cases. This approach is especially important when documents are blurry, multilingual, or partially redacted. It also helps teams build trust in the extracted output because users learn where the system is highly reliable and where it is intentionally conservative.
Why metadata enables downstream automation
Once metadata is normalized, you can automate routing, filing, and alerts. A document tagged as “referral” can be sent to scheduling. A document marked “urgent discharge summary” can be surfaced to care coordinators. A form missing a signature can generate a follow-up task without a human searching for the issue. This is where records management becomes workflow automation rather than simple storage.
Metadata also powers compliance and audit trails. You can report on which documents were received, how quickly they were processed, and which ones required manual intervention. That makes it easier to identify bottlenecks and justify process improvements. If you want examples of how structured systems generate business value beyond the obvious, see data-driven decision making.
5. A practical comparison of workflow options
The right architecture depends on document volume, sensitivity, and user experience goals. Some teams need the fastest possible upload-to-summary path, while others prioritize review accuracy and auditability. The table below compares common workflow options so you can choose a pattern that fits your use case.
| Workflow option | Best for | Strengths | Tradeoffs |
|---|---|---|---|
| Manual review only | Very low volume, ad hoc intake | Simple to implement, no integration required | Slow, expensive, error-prone, not scalable |
| OCR + human QA | High-sensitivity records, compliance-heavy teams | Strong accuracy, easy to audit | Slower than full automation, requires staffing |
| OCR + metadata extraction | Portal uploads, records management, routing | Searchable records, structured fields, automation ready | Needs tuning for document variety and edge cases |
| OCR + extraction + AI summarization | Internal teams and patient-facing apps | Fast comprehension, improved UX, better triage | Summaries must be constrained and reviewed |
| Hybrid pipeline with role-based access | Large provider networks and multi-team operations | Strong governance, flexible routing, shared source of truth | More complex implementation and policy design |
In most healthcare organizations, the fourth or fifth option delivers the best long-term return because it combines searchability, automation, and usability. Manual-only workflows may work temporarily, but they do not scale well when document volume rises. If you need a reminder of how to build robust digital systems with reliability in mind, the benchmark discipline described in secure cloud data pipelines is directly relevant.
6. Security, privacy, and compliance must be designed in
Protect the document before you enrich it
Healthcare documents often contain names, addresses, dates of birth, member IDs, diagnosis information, and treatment notes. That means the ingestion pipeline should assume sensitive data from the first millisecond. Encryption in transit and at rest is baseline. Access control, audit logging, and least-privilege permissions are equally important because OCR output and summaries can be as sensitive as the original file.
One mistake teams make is protecting storage while leaving processing layers too open. If OCR happens in one service, indexing in another, and summarization in a third, each hop must be secured. Secrets management, scoped API keys, and tenant isolation should be standard. A privacy-first OCR provider helps reduce the burden of building and maintaining those controls yourself.
Keep source, extracted text, and summary separate
Do not collapse all document representations into one undifferentiated object. Store the original file, the extracted text, structured fields, and the generated summary as separate artifacts linked by a shared document ID. This makes it easier to manage retention, reprocess documents when models improve, and show users exactly where a summary came from. It also limits the risk of a bad summary being mistaken for source truth.
Separation also helps with policy enforcement. For example, an internal team may have permission to access extracted metadata but not the original clinical attachment. A patient-facing app may show a summary without exposing full raw text. These distinctions are easier to enforce when your data model is explicit and layered.
Auditability and retention are not optional
Healthcare workflows need answerable systems, not just accurate ones. That means logging who uploaded the file, who viewed it, what was extracted, which model or OCR version produced the output, and when the record was changed. Audit trails make it possible to investigate errors and satisfy internal governance requirements. They also help you compare extraction performance across document types over time.
Retention policies should be documented and enforced automatically. Some artifacts may need to be retained as part of the medical record, while others, such as transient processing images, should be deleted quickly. Work with compliance and legal teams to define those boundaries early. For a useful reminder of how even consumer tools are under scrutiny when handling health data, review the BBC’s reporting on OpenAI’s health feature and the privacy concerns it raised.
7. How AI summarization improves usability without replacing records
Summaries speed up triage and reduce reading time
AI summarization is most valuable when staff need to understand a document quickly before deciding what to do next. A good summary can highlight the document category, key dates, named entities, and action items in a few lines. That is especially useful for intake teams handling dozens or hundreds of patient uploads each day. It reduces cognitive load and shortens time to action.
For patient-facing apps, summaries can turn a dense stack of PDFs into an understandable timeline. Instead of asking a patient to decode a 12-page faxed report, the app can say, “This document appears to be a referral letter from Dr. Lee requesting a follow-up visit within 14 days.” That phrasing is helpful, but it should always be linked back to the source file so the patient can verify the details. Summaries should guide comprehension, not override the record.
Use constrained summarization, not open-ended generation
In healthcare, uncontrolled summaries are a liability. The model should summarize only what is present in the document, ideally with guardrails around medical terminology and no speculative language. Prompting should instruct the model to avoid diagnosis, treatment advice, or unsupported inference. If the input is incomplete or ambiguous, the summary should explicitly say so rather than guess.
A safe design pattern is extract first, summarize second, and display third. Each step can be monitored and tested independently. That separation also makes it easier to improve the model later without changing your ingestion logic. For teams building human-centered automation, human-in-the-loop principles translate well to healthcare content workflows.
Summaries can power workflow automations
Beyond user experience, summaries can drive automation rules. If a summary indicates a missing signature, create a task. If it identifies an urgent referral, elevate the queue priority. If it mentions medication changes, flag the document for care-team review. These automations reduce manual interpretation and help important files move faster.
In practice, the best systems combine structured fields and summary text. Structured fields are reliable for routing, while summaries are useful for context. Together they create a highly searchable, more actionable record than either approach alone. That is the foundation of intelligent records management.
8. Implementation blueprint for developers and IT teams
Choose a document model before building the UI
Start by defining what a document object looks like in your system. Include the original file reference, OCR text, extracted entities, confidence scores, summary, status, version, and audit metadata. If your system supports multiple tenants or departments, include tenant and role scopes from day one. A clear data model prevents downstream integration pain and makes testing much easier.
Do not let the user interface dictate the data architecture. Instead, define the workflow around the record lifecycle: uploaded, processed, reviewed, approved, indexed, and archived. Each state should be observable and recoverable. That pattern reduces hidden failures and gives support teams a clear place to look when something goes wrong.
Design for retries, versioning, and reprocessing
Healthcare records are not static, and your workflow should not be either. OCR engines improve, extraction rules change, and document templates evolve. Build in the ability to reprocess a document with a newer engine or updated rules while preserving version history. That way, you can improve accuracy without losing traceability.
Retries are also essential for reliability. If a file fails due to a temporary API issue or malformed page, the system should not silently drop it. Queue-based processing, idempotent job design, and dead-letter handling are the normal ingredients of a resilient pipeline. If you are modernizing the stack around this kind of resilience, the principles in document workflow transformation are worth applying directly.
Expose document intelligence through APIs and search endpoints
Once data is structured, surface it through API endpoints that support internal systems and external apps. Common endpoints include upload, status, extracted fields, search, summary, and review actions. Search should support both metadata filters and full-text queries so developers can build flexible experiences on top of the same document index. If you support lightweight link-based processing, integration becomes far easier for web apps, CMSs, and internal portals.
For IT teams, a strong API contract is just as important as OCR quality. The result should be predictable, stable, and documented. Without that, the workflow may work in demos but fail in production integrations. Good developer experience is an operational requirement, not a luxury.
9. Real-world use cases that benefit immediately
Referral management and specialty triage
Specialty practices receive large volumes of referral documents that often need quick classification. OCR and metadata extraction can identify provider names, urgency language, relevant diagnoses, and missing attachments. Once indexed, schedulers and coordinators can search by specialty, date, or referring clinic. This shortens time from upload to appointment booking and reduces lost referrals.
The same system can flag incomplete submissions before they reach a human reviewer. For example, if the referral mentions imaging but no imaging report is attached, the workflow can trigger a follow-up request automatically. That is a small operational change with a big impact on throughput. It is also one of the clearest demonstrations of document intelligence turning into measurable efficiency.
Patient onboarding and intake automation
New patient packets often include insurance cards, demographic forms, medication lists, and consent forms. A structured workflow can classify each page, extract the needed fields, and populate downstream forms automatically. This reduces keying errors and makes onboarding faster for both staff and patients. It also improves the first impression of the digital experience.
Patient-facing apps can go one step further by showing a friendly summary of what was received and what is still missing. That kind of transparency reduces back-and-forth and makes the process feel less like document submission and more like guided onboarding. If your team is focused on user experience and conversion, lessons from structured profile optimization are surprisingly transferable: clear structure drives better outcomes.
Legacy record digitization and search
Many healthcare organizations still maintain legacy scanned records that are technically digitized but not searchable. OCR can convert those archives into usable intelligence, unlocking historical data for research, audits, and care continuity. Indexed legacy files can be searched alongside new uploads, giving staff a unified view of the patient record. That is often one of the highest-ROI document projects because the value compounds over time.
This is also where batch processing becomes especially valuable. Large backlogs can be processed in phases, starting with high-priority patients or document types. If you need to think carefully about capacity planning and infrastructure costs, the same logic used in edge compute planning and cloud cost benchmarks can help you set practical throughput targets.
10. The operating model: accuracy, governance, and continuous improvement
Measure the right KPIs
Do not measure OCR success only by raw character accuracy. In healthcare workflows, you need metrics that reflect business outcomes: time to triage, percentage of auto-routed documents, review queue size, indexing latency, and manual correction rate. You should also track document-type-specific performance because a clean typed PDF and a skewed fax have very different failure modes. These metrics tell you whether the system is improving operationally, not just technically.
Another useful metric is search success rate, or how often users find the right document on the first try. If indexing is working, this number should climb. If it is not, the issue may be document classification, metadata quality, or search relevance tuning. In other words, the output of OCR should be judged by how well it supports the workflow, not just by how good the text looks in a test panel.
Build a feedback loop for edge cases
Every healthcare operation has a long tail of edge cases: faxes with stamps, partial pages, multilingual consent forms, or poor-quality scans from old devices. Capture those failures and feed them back into document rules, review thresholds, and extraction prompts. Over time, the system becomes more accurate because it is tuned to the documents your organization actually receives. That is much more effective than chasing generic benchmark scores.
Feedback should come from both staff and end users. Staff can tell you which fields are most error-prone, while patients can report whether summaries are understandable and useful. This dual feedback loop helps you keep the workflow accurate and human-friendly at the same time. It also creates a culture of continuous improvement around records management.
Plan for governance changes over time
Healthcare data policies evolve, and your workflow should be able to evolve with them. Keep configuration externalized where possible, including retention rules, redaction policies, routing logic, and model selection. That allows compliance teams to make changes without forcing a complete code rewrite. It also reduces the risk of brittle logic buried in application code.
Over the long term, the organizations that win are the ones that treat document intelligence as a core platform capability rather than a one-off project. They have clear APIs, a stable indexing layer, and a governance model that can survive audits and scale with demand. That is what turns patient portal PDFs into searchable intelligence rather than just another pile of digital paper.
Frequently asked questions
How accurate does OCR need to be for healthcare workflows?
Accuracy should be high enough to support the specific use case, not just the document itself. For routing, classification, and search, you can often tolerate modest OCR noise if metadata extraction is strong and users can verify the source. For fields like patient names, dates, and medication details, accuracy needs to be much higher, and human review may still be appropriate for low-confidence cases. The safest approach is to define thresholds per document type and field.
Should AI summarization be used on medical records?
Yes, but only as a controlled layer on top of reliable extraction. Summaries should reflect the source document, avoid diagnosis or treatment advice, and clearly link back to the original file. In patient-facing experiences, summaries are best used to improve comprehension, not replace the record. Internal teams can use them to speed triage and identify next actions.
What is the difference between document indexing and metadata extraction?
Metadata extraction identifies structured fields such as document type, date, provider, or patient ID. Document indexing stores the extracted text and metadata in a searchable system so users can retrieve documents later. Extraction creates the data; indexing makes it useful. Most healthcare workflows need both.
How do we keep patient data private during OCR and summarization?
Use encryption, access controls, audit logs, and data separation between original files, extracted text, and summaries. Keep processing scoped to the smallest necessary trust boundary and delete transient artifacts quickly. If a third-party OCR or AI service is involved, review storage, retention, and training policies carefully. Privacy-first architecture is not just a compliance issue; it is a trust issue.
Can this workflow handle handwritten forms and faxed PDFs?
Yes, but you should expect lower accuracy than with clean typed documents. The workflow should detect difficult inputs and route them through confidence thresholds and human review. Combining OCR with page-quality checks and document-type-specific rules usually improves results. For very important fields, field validation and reviewer confirmation are still recommended.
What should we index first if we are starting small?
Start with document type, patient identifier, upload date, source channel, and a full-text OCR index. Those five elements are enough to deliver immediate searchability and routing value. Once that foundation is stable, add specialty-specific fields, confidence scores, and summary snippets. Incremental rollout reduces risk and helps teams prove value quickly.
Related Reading
- Unleashing the Power of Agentic AI in Digital Transformation of Document Workflows - A deeper look at automation patterns that can extend healthcare intake pipelines.
- Secure Cloud Data Pipelines: A Practical Cost, Speed, and Reliability Benchmark - Useful when evaluating throughput, latency, and cost for document processing.
- Edge AI for DevOps: When to Move Compute Out of the Cloud - Helps teams decide whether sensitive pre-processing should happen closer to the source.
- Agentic-Native SaaS: What IT Teams Can Learn from AI-Run Operations - A useful lens for building reliable, automated operational systems.
- The Dark Side of AI: Managing Risks from Grok on Social Platforms - A reminder that governance and guardrails matter whenever AI touches sensitive data.
Related Topics
Daniel Mercer
Senior SEO Content Strategist
Senior editor and content strategist. Writing about technology, design, and the future of digital media. Follow along for deep dives into the industry's moving parts.
Up Next
More stories handpicked for you
OCR for Research Intelligence Teams: Turning Analyst PDFs into Reusable Internal Knowledge
Designing OCR Pipelines for Financial and Market Documents That Must Ignore Cookie Banners, Boilerplate, and Duplicate Noise
Benchmarking OCR Accuracy on Medical Records: Forms, Scans, and Handwritten Notes
How to Turn Market Research PDFs into Structured, Audit-Ready Intelligence Without Breaking Compliance
Handling Compliance-Heavy Documents: Privacy Notices, Cookie Policies, and Regulatory Sections
From Our Network
Trending stories across our publication group