Innoviz + Vueron: Faster LiDAR Perception Development With Automated Annotation and a Cloud Workflow
Innoviz Technologies and Vueron have partnered to plug InnovizTwo and InnovizSMART LiDAR data directly into Vueron's VueX platform. The result: an end-to-end cloud environment that automates annotation, streamlines training and validation, and shortens the loop from raw point clouds to production-ready perception.
The joint solution is live and on display at CES 2026 (LVCC, West Hall #3569) and targets both automotive programs and smart infrastructure deployments.
What's new
- Native support in VueX for InnovizTwo (automotive) and InnovizSMART (infrastructure) LiDAR data.
- Automated LiDAR point cloud annotation with 3D bounding boxes and object classes, plus human-in-the-loop editing for accuracy.
- A single cloud environment for data storage, visualization, processing, labeling, training, validation, and deployment.
- Builds on earlier work where InnovizOne was used with Vueron's autonomous driving platform, now upgraded for higher performance sensors and broader use cases.
Why it matters for engineering teams
- Reduces manual labeling effort on 3D point clouds-typically one of the most time-consuming steps in perception development.
- Creates a consistent pipeline across automotive and infrastructure projects, so teams can reuse tools and practices.
- Supports safety-critical workflows by pairing automated annotation with expert review and fine-tuning.
- Shortens iteration cycles from data ingest to model updates, helping teams move from pilot to deployment with fewer handoffs.
How the workflow comes together
- Ingest: Capture InnovizTwo or InnovizSMART data and upload directly to VueX's cloud.
- Label: Auto-annotation generates boxes and classes; reviewers spot-check, correct edge cases, and enforce guidelines.
- Train: Launch training jobs and track metrics by dataset, scenario, or route segment.
- Validate: Run targeted evaluations on corner cases and safety-critical scenarios before promoting models.
- Deploy: Package models and push to your target stack across vehicles or roadside units.
Where this fits
- Automotive: ADAS and autonomous perception stacks that need high-fidelity 3D data and repeatable training loops.
- Smart infrastructure: Traffic analytics, incident detection, and V2I support using fixed LiDAR installations.
Live demo at CES 2026
See the integrated stack in action at Vueron's booth (LVCC, West Hall #3569) on January 6-9, 2026. If you're evaluating new perception tooling for 2026 roadmaps, it's worth a hands-on look.
Who's involved
- Innoviz: Tier-1 supplier of automotive-grade LiDAR and perception software for leading OEMs and industrial users.
- Vueron: AI-driven perception and data infrastructure company focused on turning raw LiDAR into production-ready output.
Next steps
- Explore sensor and platform details on Innoviz's site: innoviz.tech
- Plan a trial dataset flow: pick a representative driving route or junction, run auto-annotation, review accuracy, and benchmark training time vs. your current baseline.
Upskill your team
If you're building skills in perception, ML pipelines, or MLOps, browse curated learning paths by skill: Complete AI Training - Courses by Skill.
Your membership also unlocks: