@gradio-hf.bsky.social, as I transition from my full-body solution to a hands-only solution.
@gradio-hf.bsky.social, as I transition from my full-body solution to a hands-only solution.
TLDR:
Inputs
- 2-N Synchronized Videos in a zip file
Outputs
- Multiview Consistent Depthmaps + Pointclouds (Thanks VGGT + MogeV2!)
- Depth Confidence Values
TLDR:
Inputs
- 2-N Synchronized Videos in a zip file
Outputs
- Multiview Consistent Depthmaps + Pointclouds (Thanks VGGT + MogeV2!)
- Depth Confidence Values
After a refactoring, my entire egocentric/exocentric pipeline is now modular. One codebase handles different sensor layouts and outputs a unified, multimodal timeseries file that you can open in Rerun.
After a refactoring, my entire egocentric/exocentric pipeline is now modular. One codebase handles different sensor layouts and outputs a unified, multimodal timeseries file that you can open in Rerun.
I still need to add synchronization between the exo/ego videos, but all of the work the past few months is coming together
This shows a synchronization app I'm building with @rerun.io and @gradio-hf.bsky.social . More to come
I still need to add synchronization between the exo/ego videos, but all of the work the past few months is coming together
This shows a synchronization app I'm building with @rerun.io and @gradio-hf.bsky.social . More to come
Right now, such sparse input doesn't produce great results, but the depths are not being used, which should be able to help
Right now, such sparse input doesn't produce great results, but the depths are not being used, which should be able to help
Lerobot to understand better how I can get a real robot to work on my custom dataset. Using @rerun.io to visualize
code: github.com/rerun-io/pi0...
Lerobot to understand better how I can get a real robot to work on my custom dataset. Using @rerun.io to visualize
code: github.com/rerun-io/pi0...