Inside the Pipeline: How Document Intelligence Processes a Spec Sheet
A step-by-step walkthrough of a hydraulic valve spec sheet through five pipeline stages — extraction, structuring, audit, translation, and validation.
A European hydraulic components distributor receives a 4-page cartridge valve data sheet from a German manufacturer. They need it in English, French, and Italian for three EU markets. The source is a PDF with dense specification tables, a pressure-flow curve diagram, port configuration symbols, and ordering information.
Instead of sending it to a translation agency, they upload it to a document intelligence pipeline. What follows is five stages that each build on the last — and the output isn’t just a translated document. It’s structured, verified, machine-readable data available in any format.
This is a realistic walkthrough based on patterns that appear across thousands of industrial spec sheets. The document, company names, and values are illustrative, but the problems and solutions are real.
Stage 1: Extract — reading the document as data
The PDF contains four pages: product identification and specifications (page 1), performance curves and port diagrams (page 2), application data with operating conditions (page 3), and ordering information with a compliance matrix (page 4).
A generic PDF extractor would rip the text layer and produce a flat string. But spec sheet tables are spatial — the meaning of “350” depends on which column it’s in and which row header it belongs to. Position-aware extraction reconstructs table rows from Y-coordinates and columns from X-axis gaps. It also detects vector-drawn symbols — filled circles (●) for standard features and open circles (○) for optional features — from the PDF drawing operators, since these aren’t text characters and won’t appear in a standard text extraction.
What the extraction catches
- Qualifiers preserved: “Betriebsdruck: 350 bar (max. 400 bar kurzzeitig)” — the parenthetical changes the engineering meaning entirely. A naive extraction might capture “350 bar” and drop the rest.
- Multi-column table structure: A test results table with columns for Test / Standard / Requirement / Result is captured as four linked values per row, not flattened into a single string.
- Symbol table decoded: The port configuration matrix uses vector-drawn filled/empty circles. These are read from the PDF operators and mapped to the correct feature rows.
- Diagram extracted: The pressure-flow curve on page 2 is an embedded JPEG. It’s extracted, uploaded to storage, and attached to the document with a descriptive hint matched from the surrounding text.
Stage 2: Structure — from visual layout to canonical data
The extraction produced raw text with spatial awareness. Now the structuring step turns it into canonical JSON with typed property-value pairs. The AI receives both the visual PDF pages and the position-aware text layer, and outputs a structured object.
Two things happen automatically during structuring:
Domain detection
The system identifies this as a hydraulics_pneumatics document. This label isn’t just metadata — it flows downstream to the translation step, where it determines terminology choices. “Druckbegrenzungsventil” translates to “pressure relief valve” in English and “limiteur de pression” in French — not “pressure limiting valve” or “valve de soulagement de pression,” which are literal translations that a hydraulic engineer wouldn’t use.
Source language detection
The document is identified as German (DE). This matters because the decimal separator convention is different (3,8 means 3.8, not three thousand eight hundred), and compound nouns like “Druckbegrenzungsventil” need to be kept as single words, not split.
The structured output is a JSON object with sections (Technische Daten, Betriebsbedingungen, Bestellinformationen) each containing property-value pairs. Product codes, model numbers, and standard references are tagged as non-translatable.
Stage 3: Audit — finding what the source document got wrong
This is the stage that separates document intelligence from everything else. The audit compares the structured JSON against the original source document, checking every extracted value. In this case, it finds three issues.
Finding 1: Source conflict (severity: high)
Operating pressure listed as 350 bar in Technische Daten (page 1) but 315 bar in Bestellinformationen (page 4).
Both values captured. Suggested action: review with manufacturer.
This isn’t an extraction error — the source document contradicts itself. The manufacturer likely updated the specs table but forgot the ordering section. A translator would faithfully render both values in three languages, spreading the inconsistency. The audit surfaces it before that happens. For more on what automated auditing catches, see our audit guide.
Finding 2: Missing data (severity: medium)
Filtration requirement noted in the application text on page 3 (“Filterfeinheit ≤ 10 μm, β10 ≥ 75, ISO 4406: 17/15/12”) was not captured as a structured field.
Source location: paragraph below operating conditions table. Suggested action: include as specification.
The filtration requirement was embedded in prose, not in a table row, so the structuring step treated it as descriptive text rather than a specification. The audit flags the gap. The user can acknowledge it or request it be added.
Finding 3: Coverage summary
42 fields extracted across 6 sections. Source coverage: 94%. 2 diagrams attached.
The audit provides a completeness score comparing extracted data points against what it found in the source. 94% means the extraction captured nearly everything — the 6% gap is the filtration requirement (above) and two footnotes with ordering disclaimers.
Stage 4: Translate — domain-aware, not literal
The user requests translations into English, French, and Italian. The translation step works from the structured JSON, not from the PDF. It knows this is a hydraulics_pneumatics document, so it uses industry-standard terminology.
What the translation handles:
- Property names translated, values preserved. “Betriebsdruck: 350 bar” becomes “Pression de service: 350 bar” in French — the descriptive label changes, the number and unit stay untouched.
- Domain terminology, not calques. “Druckbegrenzungsventil” becomes “limiteur de pression” (FR), “pressure relief valve” (EN), and “valvola limitatrice di pressione” (IT). A generic translator might produce “valve de soulagement de pression” in French — technically comprehensible but immediately recognizable as non-specialist.
- Non-translatable items preserved. Product codes (DBDS 10 K1X/315), standard references (ISO 4406, DIN 24340), and classification codes pass through unchanged.
- Section headings translated. “Technische Daten” → “Technical Data” / “Données techniques” / “Dati tecnici.” The reader expects these in their language.
Stage 5: Validate — grammar, diacritics, and cross-language contamination
Each translated output goes through a grammar validation pass specific to its target language. The Italian output gets checked for article-noun agreement (il/lo/la) and preposition contractions (del/dello/della). The French output gets checked for accent placement and non-breaking spaces before colons and semicolons.
But the validation step also catches a subtler problem: cross-language contamination. AI translation models sometimes blend forms from related languages. Polish and Czech share many roots but have distinct endings and diacritics — “proszkowa” (Polish for powder-coated) vs. “prášková” (Czech). Spanish and Portuguese are close siblings: “-ción” (Spanish) vs. “-ção” (Portuguese). Croatian and Slovenian share South Slavic roots but have different case systems and diacritics.
The validator has language-specific rules that flag words or morphological forms that don’t belong in the target language. If a Czech adjective ending slips into a Polish translation, it’s caught and corrected here — not shipped to the customer.
The output: one extraction, multiple formats
After five stages, the same structured data serves four different people at the distributor:
| Persona | Format | What they do with it |
|---|---|---|
| Sales engineer | PDF (branded template) | Attaches to tender proposals for French and Italian customers. Logo, professional layout, print-ready. |
| Product data manager | JSON | Imports into PIM system. Structured property-value pairs feed the e-commerce catalog and product configurator. |
| Procurement team | XLSX | Downloads this valve alongside 7 competing products. Filters by operating pressure, flow rate, and seal material in a single spreadsheet. |
| QA engineer | PDF with audit report | Reviews the source conflict finding. Contacts the manufacturer to clarify the 350 vs 315 bar discrepancy before the spec sheet goes live. |
Four outputs from one upload. A translation agency would have produced three translated PDFs and none of the rest. The source conflict on page 1 vs. page 4 would have shipped to market undetected. The PIM import would require manual data entry. The procurement comparison would mean opening each PDF separately and copying values into a spreadsheet.
The same pipeline across verticals
This walkthrough used a hydraulic valve, but the pipeline works the same way on any structured technical document. The domain detection adapts terminology and audit focus:
- Coatings TDS: Drying schedule tables with temperature-dependent values, film thickness ranges (DFT/WFT), VOC content, surface preparation requirements per ISO 12944. The audit checks that recommended DFT values don’t conflict between the Application Data and Quality Control sections.
- Construction materials: EN classification codes (C2TES1 for tile adhesives), mixing ratios, coverage rates by trowel notch size. The audit flags when a product references a superseded EN standard version.
- Food processing equipment: Stainless steel grades (304 vs 316L), surface finish Ra values, IP ratings, FDA/EC 1935/2004 food contact compliance. The domain detection ensures “CIP” stays as the English loan term in languages where it’s standard usage.
- Electrical/electronics: IP protection ratings, IEC standards, operating temperature ranges, pin configurations. Compliance matrices with certification marks are decoded from vector graphics.
The pipeline is domain-agnostic by design. It doesn’t hardcode terminology for any single industry — the domain label detected in Stage 2 dynamically guides every downstream decision.
What this means for your workflow
If you process technical documents regularly — receiving spec sheets from suppliers, preparing documentation for new markets, maintaining product catalogs, or feeding data into PIM and ERP systems — the pipeline described here replaces several manual steps that are currently spread across different tools and teams:
- Manual data entry from PDFs into spreadsheets or PIM systems
- Translation agency management, DTP rework, and review cycles
- Quality checks comparing translated versions against source
- Reformatting for different output needs (branded PDF, raw data, comparison spreadsheet)
Document intelligence collapses these into a single upload. The output is structured, verified, and available in any format. Translation is there when you need it — but it’s one capability of the pipeline, not the only one.
For a deeper comparison of how this differs from traditional spec sheet translation, read translation vs. document intelligence. To understand the cost implications of each approach, see our pricing analysis.