RunLocal AI logo

RunLocal AI

The On-Device AI Development Platform

Summer 2024active2024Website
Artificial IntelligenceDeveloper ToolsML
Sponsored
Documenso logo

Documenso

Open source e-signing

The open source DocuSign alternative. Beautiful, modern, and built for developers.

Learn more →
?

Your Company Here

Sponsor slot available

Want to be listed as a sponsor? Reach thousands of founders and developers.

Report from about 2 months ago

What do they actually do

RunLocal AI is building a development tool to help engineering teams get AI models running efficiently on phones and edge devices. The product is currently in private beta; the company notes it pivoted from a previous product and is offering the new tool via demo only runlocal.ai YC profile.

In demos, teams pick a model and target device (e.g., Qualcomm, Nvidia Jetson), then use RunLocal’s agent—described as fine-tuned to chip-vendor SDKs—to port and optimize for latency and memory on-device. The tool records experiments and successful steps so teams can reuse what worked and shorten future optimization cycles runlocal.ai.

Who are their target customer(s)

  • Mobile app ML engineers building on-device features: They need models to run fast and reliably across diverse phone chips and spend time chasing device-specific crashes and slowdowns.
  • Edge and robotics engineers deploying vision or speech on embedded boards: They must meet strict latency and power budgets but lack a repeatable way to port and tune models for each hardware target.
  • Platform engineers supporting multiple hardware vendors: They manage ongoing vendor-specific fixes and performance patches that break with OS or chip updates.
  • ML Ops / production ML engineers: They deal with long, manual optimization cycles and no centralized record of what tuning steps worked, slowing releases and onboarding.
  • Small startups or R&D teams without hardware-optimization experts: They cannot afford lengthy specialist-led tuning and need faster, lower-risk routes to acceptable on-device performance.

How would they acquire their first 10, 50, and 100 customers

  • First 10: Use YC/founder intros and direct outreach to teams already shipping on-device ML; run hands-on pilots that deliver concrete optimizations and a documented result, then secure named references runlocal.ai YC profile.
  • First 50: Leverage early references to target similar companies with a standard pilot playbook and light assistance from a RunLocal engineer; co-host technical demos with chip-vendor programs to reach lookalike teams.
  • First 100: Productize onboarding with checklists and a public “what a pilot looks like” pack to reduce hand-holding; add vendor marketplace listings, small channel/sales engineering capacity, and a referral incentive to source steady pilots.

What is the rough total addressable market

Top-down context:

Analysts estimate the edge AI software market at about $1.95B in 2024 with a projection of ~$8.9B by 2030, and the broader on-device AI market (hardware+software) around $10.8B in 2025 growing rapidly Grand View Research Grand View Research.

Bottom-up calculation:

If 3,000–5,000 teams worldwide actively deploy on-device ML (mobile apps, robotics, embedded) and a specialized tool captures $25k–$60k ARR per team, the near-term SAM is roughly $75M–$300M; mid-term expansion grows with broader hardware/SDK coverage and multi-team enterprise deals.

Assumptions:

  • 3,000–5,000 active teams doing on-device ML across mobile, robotics, and embedded today.
  • Typical contract value $25k–$60k ARR per team for optimization/experimentation tooling.
  • Initial focus on software tooling budget (not counting hardware/services).

Who are some of their notable competitors

  • Edge Impulse: End-to-end edge ML platform for building, optimizing, and deploying models on embedded and edge devices—many teams use it to manage the on-device ML lifecycle.
  • OctoAI (formerly OctoML): Provides model optimization and inference infrastructure; originally focused on automated model acceleration and remains an option for teams seeking faster inference.
  • Deci AI: Optimization platform and inference SDK that improves model latency/throughput on various hardware—an alternative for performance tuning.
  • Intel OpenVINO: Toolkit for optimizing and deploying models on Intel CPUs/GPUs/VPUs; many teams use it directly as a vendor SDK instead of third-party platforms.
  • NVIDIA TensorRT: High-performance SDK for optimizing and running inference on NVIDIA GPUs and Jetson; a standard path for edge teams in the NVIDIA ecosystem.