Expert · Lesson 14 — AR research sprints when hardware isn't shipped
E14Expert
Expert · Lesson 14● live

AR research sprints when hardware isn’t shipped

How to ship Phase-0 R&D against unreleased hardware. The Parley approach to Everysight Maverick AI when you can’t hold the device.

20 min read · 60 min applyprereq: Expert 12 (custom agent files)

The unreleased hardware problem

Parley is building an AR-glasses product for real-time bidirectional transcription between deaf and hearing users. The canonical hardware target is Everysight Maverick AI. As of May 2026, the Maverick AI has not shipped commercially. The public spec sheet is limited. You cannot hold the device and run your pipeline against it.

This is not an unusual situation. Most serious AR development teams work against hardware that’s either not shipped or not in their hands. The question is not “how do we wait?” — it’s “what can we build now that will still be valid when the hardware ships?”

Parley’s Phase-0 approach: ship hardware-invariant research (ASR architecture benchmarking, signer-holdout methodology, failure taxonomy) while explicitly flagging hardware-dependent assumptions (on-device latency budget, HUD rendering pipeline, microphone placement) as unvalidated. The Phase-1 COMPLETE milestone as of April 26 — Notebooks 00, 01, and 02 all shipped, top architecture landing at 0.4467 ± 0.0097 on signer-holdout — demonstrates what’s possible before the device is in hand.

The signer-holdout gap — approximately 35 percentage points below the random-split Kaggle leaderboard — is the most important finding from Phase 1. It exposes the benchmark’s structural optimism. It’s hardware-invariant: the gap exists regardless of which AR glasses are eventually targeted. This kind of finding compounds. The hardware-dependent work comes later.

What you can and can’t do

The partition is the first discipline. Before any sprint planning, categorize every work item as hardware-invariant or hardware-dependent.

CategoryExamplesPublication readiness
Hardware-invariantASR model accuracy on benchmark and holdout datasets; signer generalization methodology; failure taxonomy; architecture comparison; dataset curation approachesPublish now — findings hold regardless of specific hardware
Hardware-inferredLatency budget estimated from comparable glasses; rendering pipeline designed for inferred HUD spec; audio preprocessing designed for assumed microphone placementPublish with explicit caveat — conclusions may change when hardware specs confirmed
Hardware-dependentOn-device inference timing; HUD-specific rendering optimization; real microphone placement effects on ASR; thermal throttling under sustained loadFlag for Phase 2+ — requires device in hand to measure

The community publication standard for Parley is higher than for internal research, specifically because the deaf community deserves accurate capability claims. The r/deaf subreddit is on explicit hold until a Deaf-community advisor has reviewed Phase-2 numbers. “Our Kaggle benchmark shows 95% accuracy” and “our product works 95% of the time for real conversations” are very different claims. The hardware assumptions register prevents the conflation.

Subscribers only · continued

The rest of Expert · Lesson 14 is for subscribers.

AR research sprints when hardware isn't shipped

  • Every Expert-tier lesson — diagnostic prompts, transcripts, prompt kits, full homework
  • Every research paper — methodology, figures, tables, reproducibility appendices
  • New Expert lessons + papers as they ship (quarterly cadence)
  • Foundations + Operating lessons stay free; bundles on GitHub stay free; this tier is the deep stuff

Free while the early catalog ships. Paid tier comes later — subscribe now and you’re grandfathered in.