Pablo Vela (@pablovelagomez1)

2025-12-24 | โค๏ธ 117 | ๐Ÿ” 13


Its done! I updated the @rerundotio demo to work with multiview videos. Right now its fully hooked up to the exoego-forge format.

The great thing about Sam3 is its robustness. I didnโ€™t need to create separate sessions for each camera. I simply generated a single instance, used a text prompt, and switched between cameras for an incremental update.

Along with all of this, Iโ€™m doing TSDF fusion on the provided depth maps + segmentation images to color the person over time.

SAM3 is such a gift to the opensource community, and Iโ€™m going to keep exploring what other cool things I can do with it =]

๋ฏธ๋””์–ด

video thumbnail


์ธ์šฉ ํŠธ์œ—

Pablo Vela (@pablovelagomez1)

Spent some more time with SAM 3. I really wanted to make it work for pointcloud/3d segmentation, so I did ๐Ÿ™‚. SAM 3 is really something, with camera parameters and images, 3d segmentation and labeling just became 10x easier.

Here I show the example of doing segmentation of a person, a yellow mustard bottle, and a book on the table.

It basically consists of running a prompt-only forward pass of SAM 3 on each image, then taking the provided depth maps of the dataset and generating a TSDF fused mesh, and updating the color based on the predicted segmentation mask

Now just need to extend it to video.

์›๋ณธ ํŠธ์œ—

๐ŸŽฌ ์˜์ƒ

Tags

3D LLM Dev-Tools