Zettascale logo

Zettascale

Energy efficient chips for AI

Summer 2024active2024Website
Artificial IntelligenceHardwareSemiconductors
Sponsored
Documenso logo

Documenso

Open source e-signing

The open source DocuSign alternative. Beautiful, modern, and built for developers.

Learn more →
?

Your Company Here

Sponsor slot available

Want to be listed as a sponsor? Reach thousands of founders and developers.

Report from about 2 months ago

What do they actually do

Zettascale (formerly Exa Laboratories) is designing a reconfigurable AI accelerator chip (an “XPU”) aimed at running training and inference with much lower energy use than today’s GPUs. As of now, they are in research and development: their site and litepaper describe the architecture and share simulated benchmarks, but there is no public evidence of shipped silicon, an SDK, or paying customers litepaper · YC profile.

Day to day, the team is iterating the hardware architecture in simulation, benchmarking their “Learnable Function Unit” building blocks (including comparisons against NVIDIA H100 in simulation), publishing technical notes, and hiring core hardware/firmware/software roles after a seed raise to move toward prototypes and tape‑out litepaper · progress post · YC profile. Their published efficiency figures are company simulations that don’t yet reflect independent, system‑level hardware tests litepaper.

Who are their target customer(s)

  • Hyperscale cloud and datacenter operators: Power, cooling, and rack density constraints drive up costs; they need more work per watt without disrupting existing software stacks. Zettascale positions its accelerator as an efficiency play for these environments litepaper · YC launch.
  • On‑prem enterprise AI teams (finance, pharma, large online services): High training costs and scarce GPUs slow iteration; any new hardware must reduce total cost and fit into existing data‑center workflows and tooling litepaper.
  • LLM/AI‑inference hosting companies and CDNs: Persistent inference loads create large, steady power bills and thin margins; they need higher throughput per watt to lower unit economics for each request litepaper.
  • National labs and academic HPC centers: Fixed power budgets and conservative procurement require predictable performance and verified energy savings; they will demand independent prototype validation before switching from GPUs litepaper.
  • Board vendors, OEMs, and systems integrators: They seek differentiated chips but worry about vendor maturity, delivery timelines, and ecosystem support; with Zettascale still at the simulation/R&D stage, execution risk remains until prototypes or tape‑outs appear YC profile · litepaper.

How would they acquire their first 10, 50, and 100 customers

  • First 10: Run tightly scoped NDA pilots with 8–10 early adopters (mix of hyperscalers, HPC labs, enterprise AI teams, and integrators) using FPGA/prototype boards or detailed simulations, co‑develop benchmarks and TCO studies, and convert wins into paid development contracts with options for first production orders.
  • First 50: Productize a repeatable evaluation kit (prototype boards + simple SDK + benchmark suite) with clear pricing/support and a refundable engineering fee; use pilot data and case studies to de‑risk purchases for OEMs, mid‑sized clouds, and on‑prem teams.
  • First 100: Scale through certified integrators and board vendors, list via cloud/OEM channels, and offer straightforward SLAs/support; accelerate sales with independent benchmarks and volume discounts tied to multi‑year support to convert evaluators into production customers.

What is the rough total addressable market

Top-down context:

AI semiconductors were ~$71B in 2024, with “AI accelerators in servers” at about $21B—most relevant near‑term for a new server‑side accelerator vendor Gartner. Broader definitions put AI accelerators at roughly ~$140B in 2024, expanding toward ~$440B by 2030 Mordor Intelligence.

Bottom-up calculation:

Using the conservative $21B 2024 server‑accelerator pool, illustrative share scenarios imply ~0.1% → ~$21M, 1% → ~$210M, and 5% → ~$1.05B annual revenue if Zettascale wins validated deployments Gartner.

Assumptions:

  • Focus on server‑side accelerators (excludes edge/consumer) as the realistic near‑term target.
  • Revenue scales roughly with share of the server‑accelerator market and average industry pricing/mix.
  • Share capture requires validated silicon, software/tooling, and initial hyperscaler/enterprise pilots litepaper.

Who are some of their notable competitors

  • NVIDIA: Incumbent datacenter GPU supplier (e.g., H100) with a mature software ecosystem; the default choice for large‑scale training and inference today NVIDIA H100.
  • Google (TPU family): Custom tensor ASICs (TPUs) integrated with Google Cloud, offering tightly coupled hardware/software for large‑scale training and inference Cloud TPU docs.
  • Cerebras Systems: Wafer‑scale accelerator and rack systems targeting very large models, competing on memory bandwidth and time‑to‑train Cerebras CS‑2.
  • Graphcore: IPU processors and systems with an alternative architecture and software stack for training/inference workloads Graphcore IPU.
  • SambaNova Systems: Integrated racks built around a reconfigurable dataflow processor and compiler, emphasizing model‑specific mapping and system‑level efficiency DataScale.