机器之心 JIQIZHIXIN (@jiqizhixin)

2026-02-01 | ❤️ 172 | 🔁 22 | 💬 2


What if your robot or car could see depth more clearly than a top-tier RGB-D camera?

Researchers from Ant Group present LingBot-Depth.

It treats sensor errors as “masked” clues, using visual context to intelligently fill in and refine incomplete depth maps.

It outperforms leading hardware in both precision and coverage, creating a unified understanding of color and depth for robotics and autonomous driving.

Masked Depth Modeling for Spatial Perception

Project: https://technology.robbyant.com/lingbot-depth HuggingFace: https://huggingface.co/robbyant/lingbot-depth Paper: https://github.com/Robbyant/lingbot-depth/blob/main/tech-report.pdf

Our report: https://mp.weixin.qq.com/s/_pTyp6hwmnLQUEeEzdp2Qw


📄 원문 내용

링크

Robbyant - Exploring the frontiers of embodied intelligence. We focus on foundational large models for embodied AI: LingBot-Depth (spatial perception), LingBot-VLA (vision-language-action), LingBot-World (world models), LingBot-VA (video action). Technology-driven, building an AGI platform for the physical world. 灵波科技 - 探索具身智能的上限。专注具身智能基础大模型:LingBot-Depth 空间感知、LingBot-VLA 视觉-语言-动作、LingBot-World 世界模型、LingBot-VA 视频动作。技术驱动,打造物理世界的 AGI 平台。

robbyant/lingbot-depth · Hugging Face

We’re on a journey to advance and democratize artificial intelligence through open source and open science.

lingbot-depth/tech-report.pdf at main · Robbyant/lingbot-depth · GitHub

Masked Depth Modeling for Spatial Perception. Contribute to Robbyant/lingbot-depth development by creating an account on GitHub.


미디어

image


Tags

AI-ML 3D-Vision Robotics