Guide Labs logo

Guide Labs

Interpretable AI models and agents that are easy to audit and…

Winter 2024active2024Website
Sponsored
Documenso logo

Documenso

Open source e-signing

The open source DocuSign alternative. Beautiful, modern, and built for developers.

Learn more →
?

Your Company Here

Sponsor slot available

Want to be listed as a sponsor? Reach thousands of founders and developers.

Report from 29 days ago

What do they actually do

Guide Labs builds AI systems that try to explain their own outputs. Instead of only returning an answer, their models aim to show which parts of the input mattered and which training‑data patterns influenced the result, with concrete research artifacts like Atlas (dataset concept annotation) and PRISM (training‑data attribution) described in their posts introducing Guide Labs and Interpretable Intelligence.

Today, the public, usable piece is InfEmbed, an open‑source Python library that helps find cohorts of test samples where a model tends to fail for faster error discovery and debugging InfEmbed. Access to their interpretable models is limited via a waitlist/early‑access API; interested users apply for access and, if accepted, can call an API that returns answers plus human‑readable explanations and training‑data attributions waitlist Interpretable Intelligence.

Who are their target customer(s)

  • ML researchers and model developers: They need to identify why models fail on specific inputs and which training examples cause mistakes; current methods are ad‑hoc and time‑consuming. Guide Labs’ work focuses on tracing predictions back to training‑data patterns to make debugging more systematic Interpretable Intelligence.
  • ML engineers and MLOps teams running models in production: They struggle to prioritize and fix real‑world failures because outputs are opaque and it’s hard to see which input components caused errors. Guide Labs provides a debugging library and plans API features to surface input‑level importance and failure cohorts InfEmbed.
  • Compliance, audit, and risk teams at regulated companies: They must produce auditable, defensible explanations for automated decisions, but black‑box models don’t provide traceable rationales. Guide Labs positions its models to return explanations and attribution auditors can inspect YC profile.
  • Healthcare and life‑science teams using AI for clinical or research decisions: Clinicians and researchers require traceable reasons and data sources to validate safety and reproducibility. Guide Labs highlights medicine and drug discovery as example domains where explainability is required Interpretable Intelligence.
  • Finance and lending teams making credit or fraud decisions: They face regulatory and customer requirements to explain denials or flags, yet current models don’t show which data drove a decision. Guide Labs emphasizes training‑data attribution and input‑level explanations for traceable decision rationale YC profile.

How would they acquire their first 10, 50, and 100 customers

  • First 10: Identify and invite heavy InfEmbed users and researchers working on interpretability for hands‑on pilots and rapid case studies; reach out via GitHub, direct emails to relevant authors, and the existing waitlist channel InfEmbed waitlist.
  • First 50: Run short, scoped pilots with ML/MLOps teams at startups and mid‑market companies in regulated verticals, provide integration support and clear remediation workflows, and convert via referrals and early case studies plus YC/research networks and talks/workshops YC profile Interpretable Intelligence.
  • First 100: Turn pilot wins into formal engagements with vetted audit workflows, compliance documentation, and onboarding for legal/risk, while building integrations with common MLOps stacks; acquire through partnerships, consultancy referrals, published playbooks/webinars, and a public API/onboarding flow fed by the waitlist waitlist introducing Guide Labs.

What is the rough total addressable market

Top-down context:

Guide Labs sits in explainable AI and AI governance, adjacent to MLOps/model monitoring. Published estimates put explainable AI at ~USD 7.8B in 2024 growing to ~USD 21.1B by 2030, AI governance software at ~USD 15.8B by 2030, and MLOps in the low‑single‑digit billions today with strong growth Grand View XAI Forrester governance Grand View MLOps GMI MLOps.

Bottom-up calculation:

Conservative TAM uses the XAI category alone (~USD 7.8B today, ~USD 21B by 2030) Grand View XAI. A realistic TAM adds governance and the portion of MLOps spend tied to auditability, then applies an overlap discount, yielding roughly USD 30–40B by 2030 (synthesizing XAI ~21B + governance ~15.8B + part of MLOps, minus 40–60% overlap) Forrester governance Grand View XAI Grand View/GMI MLOps.

Assumptions:

  • Market‑report categories overlap; apply a 40–60% overlap discount when combining XAI, governance, and MLOps.
  • Only a share of MLOps/governance spend is relevant to explainability and attribution features.
  • Regulated‑vertical upside depends on policy adoption; 5–20% of vertical AI budgets may go to explainability/governance over time.

Who are some of their notable competitors

  • Fiddler AI: Enterprise AI observability with per‑prediction explanations, diagnostics, and LLM guardrails; competes on explainability/auditing needs but focuses on monitoring/guardrails rather than models that natively expose training‑data attributions Fiddler Explainable AI.
  • TruEra: Model intelligence tooling for explainability, testing, and lifecycle monitoring; overlaps on explainability/root‑cause analysis but is positioned as a diagnostic layer for existing models, not an interpretable model architecture with training‑data attribution TruEra Explainable ML monitoring.
  • Arize AI: ML/LLM observability focused on tracing, cohort analysis, and production monitoring; overlaps on failure cohort discovery and behavior tracing but centers on observability workflows rather than models that self‑explain training influences Arize.
  • Weights & Biases: Experiment tracking and MLOps with LLM tracing/debugging (Weave); competes on developer debugging and production observability, not on native interpretable model architectures or training‑data attribution APIs W&B tracking LLM tracing guide.
  • WhyLabs: AI observability and explainability stack with open‑source logging libraries; previously a commercial competitor, the company recently discontinued operations and open‑sourced much of the platform, leaving community tooling but less direct commercial competition WhyLabs docs WhyLabs note.