[D] On-Device Real-Time Visibility Restoration: Deterministic CV vs. Quantized ML Models. Looking for insights on Edge Preservation vs. Latency.

[D] On-Device Real-Time Visibility Restoration: Deterministic CV vs. Quantized ML Models. Looking for insights on Edge Preservation vs. Latency.

Hey everyone,

We have been working on a real-time camera engine for iOS that currently uses a purely deterministic Computer Vision approach to mathematically strip away extreme atmospheric interference (smog, heavy rain, murky water). Currently, it runs locally on the CPU at 1080p 30fps with zero latency and high edge preservation.

We are now looking to implement an optional ML-based engine toggle. The goal is to see if a quantized model (e.g., a lightweight U-Net or MobileNet via CoreML) can improve the structural integrity of objects in heavily degraded frames without the massive battery drain and FPS drop usually associated with on-device inference.

For those with experience in deploying real-time video processing models on edge devices, what are your thoughts on the trade-off between classical CV and ML for this specific use case? Is the leap in accuracy worth the computational overhead?

App Store link (Completely ad-free Lite version for testing the current baseline): https://apps.apple.com/us/app/clearview-cam-lite/id6760249427

We’ve linked a side-by-side technical comparison image and a baseline stress-test video below. Looking forward to any architectural feedback from the community!

submitted by /u/tknzn
[link] [comments]

Liked Liked