The Validation Layer for AI-Generated Mobile Apps.
Run one command before you submit. ~30 seconds. Four layers. The independent check between code generation and the app store.
Mobile QA was already broken.
Apple Guideline 5.1.2(i) — effective Nov 13, 2025
Any third-party AI call now requires explicit user consent and the AI provider must be named. Most AI-coded apps ship in violation on day one. No LLM with a pre-Nov-2025 training cutoff knows it exists. Compliance moves faster than retraining cycles.[5]
One command. ~30 seconds. Release-ready.
Static Compliance
Manifest/Info.plist parsing · permissions vs. usage · SDK violations · 5.1.2(i) AI disclosure · content rating.
Crash & Performance
Top 20 device + OS combos · cold-start, ANR, OOM · catches the 34% of bugs that only reproduce on real devices.[15]
Security
Static + dynamic · hardcoded secrets · insecure storage · vulnerable deps · OWASP MASVS baseline.
Functional Behavior
AI-powered exploratory testing · walks the app like a user · flags visual regressions and broken flows.
The team that's lived this problem.
Vivek Soneja
Engineering leader at PhonePe — one of India's largest fintech apps, hundreds of millions of users, billions of transactions. Previously shipped core Android at Flipkart. Brings mobile depth at regulated scale, the PhonePe design-partner relationship, and Bay Area investor network.
Rajesh Kumar
Engineering leader at Multiplier. Previously at PayPal — Settlement Hackathon winner, 2015. Brings payments and compliance-heavy systems experience, plus the ability to run engineering in India — lower burn, deep mobile talent pool.
"Which is why we know exactly what's about to break →"
Today: a release-ready signal. Year 3: the validation graph.
Three runtime layers · LLM cascade · 80% of findings resolve on cheap models.
Compliance Parsing
Manifests, plists, SDK versions, permissions. ~10K policy rules across Apple, Google, Samsung, Amazon, Huawei.
Device Matrix
Cloud farm runs the binary across the top 20 device + OS combinations. Cold-start, ANR, OOM, battery, network.
AI Behavioral
Walks the app like a user. Visual regressions, broken flows, and a paste-ready fix prompt for Cursor/Claude Code/Copilot.
Verification Cascade — the cost moat
80% of scans resolve on cheap models · $0.20–$0.40 blended costOwn the pre-submission moment
OSS CLI · GitHub Action · 10K free users · 5 design partners · PhonePe live.
PLG bottoms-up to enterprise
First enterprise AE · CI/CD integrations · 500 paid teams · multi-SDK compliance packs.
The validation graph
Cross-customer failure intelligence · SDK + device fingerprints · "CrowdStrike for app quality."
$1B SOM. Bottom-up. Conservative.
(~20% capture)
BrowserStack
Five convergent shifts. All inside 18 months.
AI codegen tipping point
80% of new GitHub devs use Copilot week-1. Acceptance rate 27–30% — more code, more bugs.[1]
Mobile codegen mainstream
Rork, RapidNative, Replit Mobile, Cursor mobile flows — all <12 mo old · 743K monthly visits and accelerating.[11]
Store enforcement tightened
Apple 1.93M[3] + Google 2.36M[4] + the new 5.1.2(i) AI rule, Nov 13 2025.[5]
Model capability threshold
Opus 4.5 and GPT-5.1 reason about compliance, SDKs, crash traces at a quality that didn't exist 18 months ago.
Maker/checker is doctrine
Anthropic · OpenAI · DeepMind all publish on independent verification. We're productizing it for one specific domain.
"The system that produces a decision is never the system that validates it."
Everyone else hands you tools. We hand you a release-ready signal.
| Player | What they sell | Why they lose to us |
|---|---|---|
| BrowserStack $381M / $4B | Cross-browser + device farm | Infrastructure, not a release-ready signal. You still write the tests. No compliance intelligence. |
| Sauce Labs | Selenium / Appium cloud | Same — IaaS. Built pre-AI. Manual test authoring required. |
| Firebase Test Lab Google | Free Robo crawler + device access | Android only. No compliance. No fix prompts. No incentive to block submissions to its own store. |
| Sofy.ai $9.6M raised | No-code mobile test automation | No round since 2022. 2019-era positioning. Vibe-coders don't author tests — they want a one-shot ready check. |
| Spur YC S24, $4.5M | AI QA for web e-commerce | Web vertical. Different buyer. Validates the category, not a competitor. |
| AuCert | A release-ready signal + fix prompt | Cross-model · pre-submission · 10K+ store rules · open-source CLI hook · enterprise compliance moat. |
Verification Cascade
$0.20–$0.40 blended cost per scan. ~75–85% gross margin. A naive wrapper burns cash.
Validation Graph
SDK + permission + device = failure. Leaving us means flying blind. CrowdStrike's exact moat.
Device + Rule Library
Kept current by a dedicated compliance research function. A team, not a prompt.
How we charge. How we acquire.
Credits-based · unlimited seats · release-driven. Clay's exact model: $0 → $100M+ ARR.
Free OSS CLI builds the install base. Paid cloud + enterprise compliance is the moat.
Pre-launch. Soft commits. $150–180K run rate.
- Working CLI: aucert validate
- LLM cascade — Haiku → Sonnet → Opus, end-to-end
- Rule pack: 5.1.2(i), iOS 17/18, top-20 Play flags
- PhonePe · $150K/yr verbal
- Beans.ai · Multiplier — design-partner convos
- 7 warm VC intros · Nishant Mittal's network
- Cloud device farm — top 20 Android + iOS
- AI behavioral testing layer
- GitHub Action · OSS CLI · Web dashboard
Year-1 ARR bridge — $150–180K run rate
USE OF FUNDS
OSS → ENTERPRISE COMPARABLES
Mobile testing alone is tens of billions in aggregate value. We're playing in a proven category — with an AI-native wedge, a regulated gate, and a network that compounds.
invest@aucert.ai · Vivek Soneja & Rajesh Kumar