← Back to Blog
Positioning

GeraLens vs. Snap Scan vs. Pinterest Lens vs. ARKit: Adjacent Projects Compared

Published 21 April 2026 · 9 min read

Coming soon — join the waitlist

Quick answer. Snap Scan commerce has the surface but locks into Snap. Pinterest Lens does discovery but not transaction. Apple ARKit is a runtime, not a commerce protocol. Open-source toolkits (MediaPipe, OpenMMLab) are recognition libraries, not commerce infrastructure. GeraLens fits in the protocol + consent + cross-vertical action layer — the part none of the above fully occupy.

Snap Scan commerce

Snap built the most aggressive camera-to-commerce product in consumer apps — point at a shirt, buy the shirt. The experience is smooth within Snap’s ecosystem.

Overlap: camera-to-commerce UX. Gap: Snap ecosystem lock-in, no published protocol, limited for local services, no consent / audit layer for non-retail verticals. Relationship: fine as a consumer-product inspiration; different target user.

Pinterest Lens

Pinterest Lens is the longest-running visual discovery surface. It shines on fashion, home decor, and food inspiration.

Overlap: visual query → relevant results. Gap: discovery, not commit. A Pinterest Lens hit surfaces ideas, not bookings. No transactional state. Relationship: complementary — a Pinterest- style board could feed GeraLens for the commit step.

Apple ARKit / RealityKit / Visual Lookup

Apple’s stack is excellent — on-device recognition, strong privacy posture, cross-app integration. Visual Lookup in iOS identifies landmarks, plants, pets, books with restraint.

Overlap: on-device recognition with privacy respect. Gap: Apple platform-bound; no cross- platform protocol; no transactional intent-resolution beyond what Apple itself builds; no cross-vertical Gera integration. Relationship: complementary. A GeraLens iOS client should use Visual Lookup where it fits and fall back to our pipeline for commit-shaped actions.

Google Lens

Covered in an earlier post. Short version: excellent identification, not commit- shaped, heavy Google-ecosystem reliance.

Open-source recognition toolkits (MediaPipe, OpenMMLab, Ultralytics)

Strong recognition libraries; no commerce glue. Relationship: these are candidate components of the Stage-2 cloud recognition service under GeraLens.

Where we are genuinely different

  1. Commit-shaped by default. Every pipeline is wired to a transactional action, not just an identification result.
  2. Consent + audit layered in. No transaction without a scoped consent; every pipeline stage logs.
  3. Cross-vertical routing. One camera activation can hit GeraEats, GeraHome, GeraSure, GeraRide depending on the descriptor.
  4. Bright-line refusals. No face recognition, no minors, no private-environment scans — in the spec, not just policy.

Where we might be wrong

  • On-device gate models add friction and battery cost.
  • Regional visual vocabulary is a coordination nightmare.
  • Users may not care about audit; product-design has to make the audit layer visible without annoying them.

Cooperative future

The right 2030 stack: Apple / Android for on-device primitives, MediaPipe-class libraries for recognition, GeraLens for protocol + consent + routing, GeraNexus for commit. Each layer swappable.

Help us design ambient discovery.

Join the waitlist