Orphic
Runlog AI

RFQ Document Inference

Extract data from unpredictable data dumps by populating a benchmark ontology which updates in real time in light of new information.

4 weeks
6 synthetic data dumps
API Integration
Schemas by Category

RFQs are Data Intensive

Every customer has a unique data representation which resists repeatable RFQ automation agnostic to differing conventions and jargon

Financial Modeling for RFQs require understanding relevant sections of bulky data

Every project has its own representation of relevant data coming from customization on the company's ground truth

Even this ground truth has drastic variation across companies

Fundamentally, this data and the manipulation thereof is informed by the same engineering principles but there is neither knowledge eagerness to adapt a universal ontology for RFQ data across organizations

Current Bottleneck

Orphic must build custom parsers and interpreters for each customer providing a unique represented data with varying focus and relevance

Atlas Solution

Atlas maintains a living universal schema for all slices of interest across spreadsheets, PDFs, and JSONs, identifying and extracting relevant fields with confidence.

Security & Data Governance

This operates on synthetic data. Atlas does not have access to any additional data.

Pilot guardrails (default)

  • Internal data from companies will not be compromised
  • No training on Enterprise data; configurable retention/deletion
  • Role-based access; audit logs for every access and export

We can tailor controls to Orphic's internal AI pilot review process; the pilot does not require deep integration.

Pilot Overview

4 weeks to understand 6 Data Dumps

Each data dump is its own project. The goal is to know what is relevant in each project.

RFQ Data Dump

Input Synthetic Data for 6 use cases

Extraction Memo

Schemas + Extractions + change log

API Integration

Downstream usage for insights

Orphic provides 6 data dumps with different RFQ use cases containing Spreadsheets, PDFs, and JSONs

Atlas converts each file into Natural Partitions with its corresponding extractions

The identified partitions and their schematized extractions are available via API integration

Schemas for Natural Partitions can enhance over time with the complete extraction history available

Impact Tracked

Primary: Schema Formation with extraction correctness

Ontology

Defined structure for speciic types of information

e.g. Knowing what a 'Grade-based-Price for X' is and identifying the relevant aspects for it across all occurrences

Baseline

Non-Existent

Target

Stable inferred schema

Extraction Confidence

Likelihood of correctness for all extractions

Review Count is the number of observations which need to be reviewed to identify and correct all mistakes

Baseline

Non-Existent

Target

95% Accuracy, 10% Review Count

Week-by-Week Structure

3 weeks to decision

Week 0

Setup + Definitions

  • Atlas provides all Data Dumps
  • Agree on pilot guardrails and data handling
  • Define export format for case packets (ticket + attachments + relevant exports)

Output: Signed-off pilot plan + data schema

Week 1

Baseline + First Run

  • Run Atlas on first data dump
  • Feedback session for necessary adjustments in output
  • Mid-pilot check-in

Output: First-run metrics + Understanding of insight gap

Week 2

Iteration + Second Run

  • Incorporate reviewer feedback (what was missing / misleading)
  • Feedback on Insights from first 3 Data Dumps
  • Finalzie scope and variation for post-pilot adoption

Output: Delta metrics + updated artifacts

Week 3

Results + Go/No-Go

  • Compile report, identify best-fit teams, define next scope if positive
  • Additional Data Dumps Processed

Output: Go / no-go decision + rollout plan

Engagement Includes

4-week pilot investment

Includes

  • Pilot setup + investigation packet schema alignment
  • Versioned schema memo + must-verify evidence set generation
  • Change log evaluation (investigation update injection or second batch)
  • Security/retention configuration + LLM provider options
  • Weekly review + iteration loop with investigation team
  • Final report + go/no-go recommendation + auditability assessment