Open-source stereo tracking architecture

Multi-View Stereo Tracking System

Geometry-first tracking
Not AI guesswork.

EdgeTrack is an open-source multi-view stereo tracking system built for robotics, VR, teleoperation, and spatial interaction. It combines RAW capture, precise timing, hardware synchronization, and host-side fusion to deliver stable and reproducible 3D motion data.

Instead of relying on opaque camera pipelines or AI-only estimation, EdgeTrack uses metric stereo geometry as its foundation. The system is modular, hardware-agnostic, and designed for deterministic real-world tracking workflows.

Modular architecturePrecise synchronizationOpen source (Apache 2.0)
EdgeTrack multi-view stereo tracking system

A demo video will be available on YouTube soon

What’s different

Born out of frustration with the limitations of typical off-the-shelf systems, EdgeTrack introduces a tracking architecture designed for determinism, robustness, transparency, and cost efficiency.

The system is built around a RAW-first capture pipeline, hardware synchronization, and host-side fusion. By avoiding ISP-processed or compressed streams, timing and processing remain predictable and fully controllable.

On the host side, CoreFusion combines multiple rigs, rejects outliers, smooths motion, and processes ROI-based stereo data to produce stable outputs such as metric 3D keypoints and structured motion signals for higher software layers.

In multi-rig setups, phase-shifted timing allows different rigs to operate in offset capture windows, reducing interference and improving temporal coverage. A short concept example is shown below:

More architectural concepts and pipeline variants are documented throughout the GitHub repositories.

Quick comparison

EdgeTrack is designed to remain transparent, modular, and predictable where many typical tracking stacks become difficult to control or extend.

Other SystemsEdgeTrack
ISP-processed or compressed streams, limited controlRAW-first pipeline + deterministic control
Single-rig focus or unstable multi-camera setupsMulti-rig capable and fusion-ready
Often limited to 15–30 FPS in practical setupsScales to high FPS with the right hardware and tuning
Timing jitter from USB, buffering, or encodingEthernet-first + hard sync + stable timing modes
Dense depth everywhere with heavy compute costROI-first processing for efficiency; dense depth optional
Depth inferred mainly by AI can be inconsistentMetric stereo geometry as foundation; AI assist optional
Closed ecosystems and vendor lock-inOpen source + modular components
Other Systems

ISP-processed or compressed streams, limited control

EdgeTrack

RAW-first pipeline + deterministic control

Other Systems

Single-rig focus or unstable multi-camera setups

EdgeTrack

Multi-rig capable and fusion-ready

Other Systems

Often limited to 15–30 FPS in practical setups

EdgeTrack

Scales to high FPS with the right hardware and tuning

Other Systems

Timing jitter from USB, buffering, or encoding

EdgeTrack

Ethernet-first + hard sync + stable timing modes

Other Systems

Dense depth everywhere with heavy compute cost

EdgeTrack

ROI-first processing for efficiency; dense depth optional

Other Systems

Depth inferred mainly by AI can be inconsistent

EdgeTrack

Metric stereo geometry as foundation; AI assist optional

Other Systems

Closed ecosystems and vendor lock-in

EdgeTrack

Open source + modular components

More benefits and detailed comparison tables are available in the documentation.

How it works

Timing → Tracking → Fusion → Your Application

TDMStrobe timing and sync

TDMStrobe

Precise timing and hardware synchronization for multi-rig systems, including triggering and phase control.

EdgeTrack stereo capture

EdgeTrack

Edge-side stereo capture with RAW-first preprocessing, full pipeline control, and ROI-based stereo processing.

CoreFusion host-side fusion

CoreFusion

Host-side fusion combining multiple stereo rigs, rejecting outliers, smoothing motion, and producing stable metric 3D outputs.

Application integration

Your application

The final output can drive robotics, VR tools, research systems, teleoperation, or higher-level interaction software.

MotionCoder gesture layer

MotionCoder (optional)

A gesture interaction layer that maps stable motion signals to commands for 3D authoring, VR tools, and structured spatial workflows.

EdgeSense AI assist

EdgeSense (optional)

AI assist for classification and confidence scoring, without replacing stereo geometry as the foundation of the system.

See the full architecture diagrams and pipeline variants in the docs.

Hardware support

EdgeTrack is designed to stay hardware-agnostic and avoid vendor lock-in. The architecture can run on widely available compute platforms and can be adapted to different performance and cost targets.

Edge capture nodes can run on compact edge hardware such as Raspberry Pi-class systems or other ARM boards, while host-side fusion can run on standard Linux or workstation hardware. This makes the system flexible for prototyping, research, industrial setups, and cost-sensitive deployments.

Recommended reference configurations, hardware notes, and integration guidance are documented in the project docs.

Technologies we love

EdgeTrack is built on a modern stack focused on performance, openness, portability, and precise control over the tracking pipeline.

Programming languages

The implementation focuses on efficient systems programming for timing-critical and performance-sensitive parts, while higher-level languages remain useful for tooling, research, and rapid iteration.

Use cases

EdgeTrack can be applied to a wide range of fields that benefit from stable, metric, multi-view motion tracking and transparent pipeline control. Its modular architecture enables many possible workflows, from creative tools and robotics to industrial systems, teleoperation, research, and spatial analysis.

Community & updates

Join the EdgeTrack community and receive updates about development progress, releases, and hardware availability.

Access community and documentation updates with a simple email login.

No spam. Magic-link login. Unsubscribe anytime.

Logo inspiration

The EdgeTrack logo was developed from the same geometric design principles as xtan, but with a stronger emphasis on grayscale structure, spatial depth, and a hollow tetrahedral form.

Instead of being drawn manually, the shape was generated programmatically with JavaScript and Babylon.js to explore geometric surfaces, lighting, and spatial form.

Note

This section is still evolving and will be refined step by step as the project grows. Thank you for your understanding and interest.