**Meta’s SAM‑3D** is a new **foundation model for 3D** vision that extends the 2D **Segment Anything Model (SAM)** into the third dimension. It actually consists of two specialized models – **SAM‑3D Objects** (for general 3D object and scene reconstruction) and **SAM‑3D Body** (for full 3D human pose and shape estimation). Both were trained on unprecedented datasets using novel training pipelines. In this post, we’ll dive into **how SAM‑3D was trained** – covering the datasets (real vs synthe...