Here's a number that should bother every healthcare data leader: roughly 80% of clinical data is unstructured. Progress notes. Radiology reports. Pathology findings. Discharge summaries. Operative notes. Nursing assessments. It's all sitting in your EHR right now, and almost none of it feeds your analytics.
For years, we've built healthcare data platforms around the structured 20% — diagnosis codes, lab values, medication lists, claims data. That's not wrong. Structured data is reliable, queryable, and well-understood. But it's also incomplete. The richest clinical context — the reasoning behind a diagnosis, the nuances a physician noticed, the social determinants mentioned in passing — lives in free text. And until recently, extracting it at scale was either impossibly expensive or hopelessly inaccurate.
That's changing. Fast.
Why Legacy NLP Failed Healthcare
Healthcare has been trying to crack unstructured data for over a decade. Tools like cTAKES, MetaMap, and early clinical NLP systems could extract medical concepts from text — medications, diagnoses, procedures — with reasonable accuracy in controlled settings. The problem was never the technology's potential. It was the gap between demo and production.
The Three Walls
Traditional clinical NLP hit three walls simultaneously:
- Vocabulary brittleness. Rule-based systems required exhaustive dictionaries. A physician writing "pt c/o sob x 3d, worse w/ exertion" would stump a system expecting "patient complains of shortness of breath for three days." Every abbreviation, every shorthand, every institution-specific convention needed a rule.
- Context collapse. Negation, hedging, and attribution are everywhere in clinical text. "No evidence of malignancy" contains the word "malignancy" but means the opposite of what a naive extractor would conclude. Legacy systems handled negation with pattern matching (NegEx and its descendants), but struggled with complex assertions like "family history of diabetes, patient denies symptoms."
- Scale economics. Getting a traditional NLP pipeline to 85% accuracy on one document type at one institution was a six-month project. Then you moved to a different hospital system with different templates, different abbreviation conventions, and different specialties — and started over.
The result: most health systems that tried clinical NLP ended up with narrow, fragile pipelines that covered a handful of use cases and required constant maintenance. The ROI rarely justified the effort beyond research settings.
LLMs Change the Extraction Economics
Large language models don't just improve on legacy NLP — they break the cost curve entirely. The same model that handles a cardiology progress note can process a radiology report, a psychiatric evaluation, and a surgical operative note without retraining. It handles abbreviations, negation, and context natively because it learned language, not just medical dictionaries.
What this means in practice:
- Zero-shot extraction. You can prompt a model to extract structured data from clinical text it's never seen before — and get usable results. Not perfect, but usable. That's a paradigm shift from "six months to train a custom model per document type."
- Semantic normalization. "Heart attack," "MI," "myocardial infarction," "STEMI," and "acute coronary event" all map to the same concept without a manually curated synonym table.
- Contextual reasoning. LLMs understand that "ruled out for PE" means the patient does not have a pulmonary embolism. They handle hedging ("cannot exclude the possibility of...") and attribution ("mother had breast cancer") with accuracy that would have required dozens of hand-crafted rules.
The shift isn't from bad NLP to good NLP. It's from NLP as a specialized project to NLP as a pipeline component — something you configure, not something you build from scratch for every use case.
Building the Pipeline: Architecture That Works
Dropping an LLM API call into your ETL and calling it a day will get you fired — or worse, get a compliance violation. Clinical unstructured data requires a purpose-built pipeline that handles the unique constraints of healthcare: PHI protection, auditability, cost management, and clinical validation.
The Four-Layer Stack
The architecture that's emerging across mature healthcare data teams follows a consistent pattern:
- Layer 1: Extraction and De-identification. Before any text hits a model, strip or tokenize PHI. This isn't optional — it's table stakes for HIPAA. Tools like Philter, Presidio, or custom regex pipelines handle the de-identification pass. The extracted text goes to the model; the PHI mapping stays in your secure environment for re-linkage.
- Layer 2: Structured Output Generation. The LLM processes de-identified text and returns structured JSON — diagnoses, medications, procedures, social determinants, clinical findings — with confidence scores and source spans (the exact text that supports each extraction). Source spans are non-negotiable. Without them, you can't audit or validate.
- Layer 3: Normalization and Coding. Raw extractions get mapped to standard terminologies — SNOMED CT, ICD-10, RxNorm, LOINC. This can be a second LLM call, a vector similarity lookup against a terminology embedding index, or a hybrid approach. The goal: every extracted concept has a standard code, not just free-text labels.
- Layer 4: Warehouse Integration. Normalized, coded extractions land in your analytics warehouse (Snowflake, BigQuery, Databricks) as structured tables that join cleanly with your existing claims, labs, and administrative data. Now your analysts can query across structured and formerly-unstructured data with standard SQL.
Cost Control Is an Architecture Problem
A 2,000-token clinical note costs roughly $0.01–0.03 to process through a frontier LLM. That sounds cheap until you multiply by millions of notes per year at a mid-size health system. The organizations doing this well use a tiered approach: fast, cheap models for routine extractions (medication lists, vital signs mentioned in text) and larger models for complex reasoning tasks (differential diagnosis extraction, clinical timeline reconstruction). Caching and batching are critical — the same radiology report template with different findings doesn't need a from-scratch analysis each time.
Use Cases That Are Working Today
This isn't theoretical. Healthcare organizations are running unstructured data pipelines in production right now, and the use cases are expanding rapidly:
- Clinical trial matching. Eligibility criteria are notoriously complex and often require information buried in clinical notes — prior treatments, specific lab trajectories, comorbidity details that don't surface in problem lists. LLM-powered extraction against clinical notes has cut screening time dramatically at several academic medical centers.
- Social determinants of health (SDOH) capture. Physicians routinely document housing instability, food insecurity, transportation barriers, and substance use in notes — but rarely enter structured SDOH codes. Extracting these mentions from free text creates a population health dataset that didn't exist before.
- Quality measure abstraction. Chart abstraction for quality reporting is one of healthcare's most tedious manual processes. Organizations are using LLMs to pre-abstract cases, reducing the human abstractor's job from reading every note to validating pre-filled extractions.
- Prior authorization evidence. Building the clinical case for a prior auth requires pulling relevant findings from across the patient's record. Automated extraction can compile the supporting evidence in seconds instead of the 20–40 minutes a nurse typically spends per case.
- Retrospective cohort identification. Researchers needing patients with specific clinical characteristics — "history of treatment-resistant depression with trials of at least two SSRIs" — can now query across narrative text, not just coded data that may be incomplete.
The HIPAA Question Everyone Asks
Can you send clinical notes to an LLM? The answer is yes — with guardrails. The path forward involves one or more of these approaches:
- De-identify first, always. The safest approach: strip PHI before any text leaves your environment. Modern de-identification pipelines are good enough for most extraction use cases. You lose some context (names, dates, locations) but gain HIPAA Safe Harbor compliance.
- BAA-covered APIs. Major cloud providers and some LLM vendors now offer BAA-covered endpoints. If you have a signed BAA, you can process PHI-containing text — but you need to ensure the data handling, logging, and retention policies meet your compliance requirements.
- On-premises or VPC-deployed models. Open-weight medical LLMs (Med-PaLM's successors, fine-tuned Llama variants, domain-specific models from the research community) can run entirely within your infrastructure. Performance is approaching API-model quality for many clinical extraction tasks.
The right approach depends on your organization's risk tolerance, volume, and existing infrastructure. Most health systems we work with use a hybrid: de-identified text to cloud APIs for complex reasoning, on-prem models for high-volume routine extractions.
What This Means for Your Data Platform
If you're building or modernizing a healthcare data platform in 2026, unstructured data processing isn't a future roadmap item — it's a current architecture decision. The organizations that treat clinical text as a first-class data source are building fundamentally richer analytics capabilities than those still ignoring 80% of their data.
The practical steps:
- Audit your text data. What document types exist in your EHR? What's the volume? What clinical information is only captured in free text?
- Start with one high-value use case. Don't try to extract everything from every document type on day one. Pick the use case with the clearest ROI — SDOH extraction, quality abstraction, or prior auth automation — and build the pipeline end to end.
- Design your warehouse schema now. Extracted concepts need a home in your data model. Plan for provenance tracking (which document, which sentence, which model version produced this extraction) and confidence scoring. Your analysts will thank you.
- Invest in validation workflows. LLM extractions are good but not infallible. Build human-in-the-loop validation into your pipeline — especially for clinical decision-making use cases. Accuracy improves as you accumulate validated examples for fine-tuning and evaluation.
The 80% of your data that's been sitting in free text isn't a problem to solve later. It's the competitive advantage you're leaving on the table right now.
Ready to Unlock Your Unstructured Data?
CV Health builds clinical NLP pipelines that turn free text into analytics-ready data — with HIPAA compliance baked in from day one.
Let's Talk Architecture