AI-Assisted RFP & Document Analysis System
LLM & NLP — Proposal Intelligence Workflow (Concept Prototype)
Built a prototype AI-assisted workflow to analyze large RFP documents (50–200 pages), extract structured requirements, and score alignment against an internal capability matrix. The focus is speed + consistency: turning unstructured PDFs into analyst-ready outputs (requirements list, risks, gaps, and an executive fit score).
Use case: 50–200 page RFPs
Output: requirements + risks + fit score
Pipeline: chunking → retrieval → extraction → scoring
Key Contributions
- Document segmentation: designed chunking rules and section-aware splitting to preserve context (scope, SLAs, compliance, pricing assumptions).
- Extraction schema: defined a structured requirement model (category, requirement text, priority, evidence span, owner, due date, risk level).
- Scoring logic: created a capability-alignment rubric (covered / partial / gap) with risk flags for non-standard terms and high-effort obligations.
Data
- Inputs: PDF RFPs (unstructured), internal capability matrix (spreadsheet/table), optional past proposal library for retrieval.
- Output formats: JSON requirements table + risk log; optional CSV export for analysts; summary page for exec review.
- Privacy: prototype uses sanitized / synthetic RFP excerpts for portfolio demonstration.
Analyst Experience
- Executive summary: overall fit score + top risks + top gaps (bid/no-bid ready).
- Requirements table: filterable list by category, owner, priority, and due date; includes citation/evidence span back to the source section.
- Gap report: highlights partial coverage and missing capabilities with recommended follow-ups for SMEs.
- Exports: CSV/JSON outputs to plug into proposal trackers and response planning templates.
How it Works
- 1) Ingest: parse PDF → detect headings/sections → normalize text and metadata (page, section, paragraph).
- 2) Chunk + retrieve: split into section-aware chunks; retrieve top relevant chunks per requirement category.
- 3) Extract: LLM generates structured requirements using a fixed JSON schema + evidence references.
- 4) Score: compare extracted requirements to capability matrix → label as covered/partial/gap + risk severity.
- 5) Review layer: analyst approves/edits high-risk items; exports final tables and summary.
Guardrails: schema validation, confidence thresholds for risk flags, and evidence-linked citations to reduce hallucinations.
Evaluation (Prototype)
- Quality checks: % of extracted requirements with evidence spans; schema-valid output rate; duplicate/overlap rate.
- Human review: sampled requirements reviewed for correctness and category placement; risk flags reviewed for false positives.
- Operational goal: reduce repetitive scanning and improve consistency across reviewers (portfolio prototype).