I design systems that transform complex data, documents, and policy contexts into actionable decisions — bridging technical rigor with real-world impact.
"My work focuses on turning fragmented data, documents, and qualitative inputs into structured, usable intelligence."
I'm a data scientist and AI systems builder who operates at the intersection of technical architecture and applied research. My work spans decision support systems, knowledge pipelines, and economic analysis — with a consistent focus on making complex information actionable.
I bring together machine learning, NLP, and systems design with deep domain knowledge in economic and policy research. I build things that are meant to be used — not just analyzed.
AI-powered pipelines that synthesize multi-source intelligence into clear, structured outputs for high-stakes decisions.
OCR, LLM extraction, and retrieval-augmented architectures that unlock value locked in unstructured documents.
Predictive models, NLP classifiers, and statistical frameworks tuned for operational and policy contexts.
Rigorous quantitative analysis connecting data-driven methods with economic theory and policy implications.
Develop and maintain production-grade analytics pipelines and agentic workflows that unify structured records with narrative text sources. Apply ML/NLP to transform qualitative inputs into decision-ready metrics and support econometric broadband impact modeling used in policy research.
Developed supervised and unsupervised machine learning models for client research engagements, engineered ETL pipelines to prepare large datasets for modeling, and implemented evaluation practices to strengthen reproducibility.
Built predictive models to analyze membership behavior and improve retention-oriented campaign strategy. Automated recurring data pipelines and integrated structured and unstructured CRM data for fuller lifecycle visibility.
An agentic opportunity intelligence system that continuously monitors, evaluates, and routes contract and grant opportunities through a stateful, human-governed pursuit workflow.
A production-grade, multi-source RAG system with corpus, document-only, and hybrid modes that delivers citation-ready, context-grounded intelligence for policy and program decisions.
A machine learning decision system that forecasts demand and outcome quality, then operationalizes targeted intervention strategies to improve participation and resource alignment.
Published mixed-methods research combining regional econometric analysis and qualitative evidence synthesis to evaluate federal broadband funding impacts and policy implications.
Designing agentic pipelines and structured reasoning workflows that support human-governed decisions in policy and program environments.
Building NLP and embedding-based systems that extract, align, and surface meaning from heterogeneous document corpora and operational data.
Using econometric modeling and mixed-methods research to evaluate federal investment outcomes and inform infrastructure policy recommendations.
Translating complex, multi-layered datasets and document-driven findings into clear, interactive deliverables for analytical and executive audiences.
A deep look at the systems, pipelines, and analyses I've built — organized by domain and designed for exploration.
An end-to-end agentic platform that ingests contract and grant opportunities, evaluates them against organizational priorities, and executes downstream actions including prioritization, routing, summarization, and analyst escalation.
Combines automated decision support with human oversight, enabling faster and more consistent pursuit workflows.
Business development teams face fragmented, high-volume opportunity streams with inconsistent structure and changing requirements.
BOOM operates as a stateful orchestration layer that:
(1) continuously ingests and monitors opportunity sources,
(2) normalizes and semantically structures incoming data,
(3) retrieves relevant internal context from capabilities, offerings, and prior work,
(4) evaluates opportunity fit using proprietary scoring logic, and
(5) conditionally triggers downstream actions such as routing to teams, generating summaries, initiating document analysis, and preparing decision support outputs.
By automating triage, prioritization, and initial analysis, BOOM allows teams to focus entirely on high-value pursuit decisions rather than screening.
A production-grade decision intelligence RAG system engineered for high-reliability retrieval, grounded generation, and operational scalability. It supports corpus retrieval, document-scoped retrieval, and hybrid retrieval across persistent and session-level sources.
Critical documents, prior work artifacts, and institutional company IP were fragmented across disconnected databases and repositories, making knowledge hard to locate at decision time. That fragmentation slowed deliverable production, weakened proposal development velocity, and increased the risk of repeating work or missing high-value evidence during response windows.
The system unifies ingestion, retrieval, and generation in a single operational pipeline:
(1) query and document inputs are routed across corpus, document-only, or hybrid modes,
(2) multi-index retrieval materializes chunk/parent/neighbor context with metadata-aware reranking and diversification, and
(3) a managed LLM layer generates citation-grounded responses with streaming, retry/backoff resiliency, concurrency controls, token-budgeted context assembly, and telemetry for auditability.
By centralizing fragmented documents and institutional IP into a retrieval-first workflow, the system reduced knowledge loss during active pursuits, improved delivery speed, and fueled more consistent decision support across recurring and ad hoc requests.
A production qualitative research database that transforms unstructured evidence into a context-aware semantic knowledge layer for analysis, synthesis, and decision support.
Critical research evidence arrived in fragmented formats and channels, including federally mandated reports, interviews, survey instruments, and scraped media references. Manual cross-source synthesis was slow, inconsistent, and difficult to operationalize at scale.
The system was designed as a single qualitative intelligence workflow:
(1) ingest and normalize heterogeneous evidence sources into a shared, schema-flexible structure,
(2) preserve provenance and metadata at the source and record level for traceability,
(3) generate embeddings and vector indexes to enable semantic retrieval alongside structured filtering, and
(4) assemble context windows with entity tags and citation-linked references for reliable synthesis and reporting.
Consolidating cross-agency qualitative data improved research continuity, strengthened evidence reuse for reporting and proposals, and supported publication-ready analysis.
An operational intelligence platform combining machine learning, predictive workflow analytics, and resource-allocation visibility to improve delivery performance in multi-project environments.
Program leadership lacked timely integrated signals for workflow health and resource pressure. This increased exposure to burnout risk, workstream fragmentation, and poor coordination across interdependent project teams.
The system was built as a unified operational decision workflow:
(1) integrate workflow, staffing, and delivery signals into a shared operations model,
(2) map dependencies between workstream stages and resource capacity to expose coordination pressure points,
(3) apply forecasting, anomaly detection, and risk scoring for burnout, fragmentation, and handoff bottlenecks, and
(4) surface role-specific dashboards and alerting layers to support earlier intervention and allocation decisions.
Teams completed workstreams in fewer hours than prior projections while improved division of labor reduced duplicate effort and clarified ownership. Operational response cycles accelerated as risks were surfaced sooner and addressed earlier.
Designed and implemented a predictive system that forecasts participation dynamics and outcome quality, then translates those forecasts into targeted intervention strategies. In federal participation-driven programs, this enables teams to anticipate variability, optimize resource allocation, and influence outcomes instead of reacting after trends emerge.
Organizations operating in uncertain, participation-driven federal environments lacked visibility into future demand and outcome distribution. This produced resource mismatches (overload vs. underutilization), ineffective or poorly timed outreach, and limited ability to correct course once trends appeared. Even where forecasts existed, there was no structured system for operational action.
The system was implemented as a closed-loop predictive decision workflow:
(1) model expected demand and outcome quality from historical program data, behavioral signals, and structural characteristics,
(2) calibrate probabilities and uncertainty ranges to support reliable deployment thresholds,
(3) diagnose key drivers with feature attribution and segment scenarios (low participation risk, low quality risk, imbalance), and
(4) activate targeted intervention strategies through decision rules that link forecasts directly to operational actions.
Enabled proactive resource allocation aligned to predicted demand, improved participation and outcome quality through targeted interventions, and reduced intuition-only decision making by embedding data-driven rules into federal program operations.
A published research project analyzing the economic impact of federal broadband funding through a custom mixed-methods framework that pairs regional econometric analysis with large-scale qualitative evidence synthesis.
Policy stakeholders needed causal, region-specific estimates of economic effects, but quantitative findings alone were insufficient for implementation interpretation. Existing analyses lacked a unified structure linking economic outcomes to on-the-ground qualitative evidence from communities and program actors.
The research used an integrated mixed-methods design:
(1) regional econometric modeling estimated effects across employment, income, and local business indicators,
(2) qualitative evidence was systematically indexed and analyzed to interpret mechanisms and implementation constraints, and
(3) both evidence streams were synthesized into a single policy analysis layer for stronger confidence, sensitivity testing, and applied guidance.
Findings informed federal broadband policy conversations and implementation strategy discussions, with integrated quantitative and qualitative evidence improving confidence in where and why impacts were strongest.
Combined listening-session NLP analysis with implementation scenario research into a single service layer for policy intelligence. The system synthesizes unstructured inputs, policy documents, and geography-linked infrastructure data to surface actionable implementation insights.
Public inputs and state implementation choices were being analyzed in silos, obscuring cross-signal patterns. Stakeholders lacked an integrated framework to connect thematic concerns, sentiment trends, geographic disparities, and likely policy outcome tradeoffs.
The system used a unified policy intelligence workflow:
(1) ingest comments, transcripts, and planning artifacts into a shared analysis layer,
(2) apply supervised and weakly supervised classification, semantic clustering, sentiment analysis, and topic modeling to surface dominant concerns and emerging issue clusters,
(3) link NLP outputs to geospatial coverage and demographic context for comparative scenario assessment, and
(4) communicate results through advanced visuals including choropleth and bivariate maps, Sankey flows, and stakeholder-topic network graphs.
Delivered integrated NLP and geospatial evidence products that improved policy scenario evaluation and accelerated synthesis during active federal and state broadband decision windows.
Open to collaborations in AI systems, data science, and applied research.
Open to contract, advisory, and collaborative research engagements in AI systems and policy analysis.
RAG architectures, decision pipelines, document intelligence
NLP, predictive modeling, analytical systems for civic/gov contexts
Broadband, digital equity, federal program analysis and evaluation
Presented research at an international, multidisciplinary conference of economists, policymakers, and regional scientists focused on spatial and economic analysis. Shared a mixed-methods framework integrating quantitative modeling with qualitative insights to inform broadband policy and regional development strategies.
Panelist at a national convening hosted by multiple Federal Reserve Banks, bringing together researchers, policymakers, and practitioners to advance work on digital access and economic inclusion. Contributed perspectives on equity-focused broadband deployment and the role of data in identifying and addressing disparities in underserved communities.
Delivered a policy-focused presentation for a national broadband policy audience, examining evolving BEAD implementation strategies and demonstrating how NLP methods can be used to systematically analyze policy documents and public input at scale.
Panel discussion with regional leaders and practitioners on practical approaches to identifying high-value AI use cases within organizations, with emphasis on implementation strategy, change management, and aligning technical capabilities with operational needs.