Meta has introduced SAM 3 and SAM 3D, two cutting‑edge models that redefine how artificial intelligence perceives, interprets, and reconstructs the world around us. These innovations extend the Segment Anything Collection, pushing boundaries in object detection, multimodal understanding, and 3D reconstruction — and they’re poised to reshape industries from creative media to commerce.

SAM 3: Precision Meets Creativity

  • Text‑Driven Segmentation: SAM 3 responds to natural language prompts, isolating objects with remarkable accuracy. A query like “red baseball cap” instantly highlights every matching item in an image or video.
  • Nuanced Queries: Integrated with multimodal LLMs, SAM 3 can handle complex instructions such as “people sitting down, but not wearing a red baseball cap.”
  • Creative Empowerment: In Meta’s Edits app, creators can apply effects to specific people or objects, while Vibes on the Meta AI app offers playful, interactive experiences.
  • Open Source Impact: Released with model weights, benchmarks, and annotation tools, SAM 3 empowers developers and researchers to fine‑tune and innovate.

SAM 3D: Breathing Life into Images

  • Single‑Image Reconstruction: SAM 3D can generate 3D objects and human body models from just one photo.
  • Dual Models:
    • SAM 3D Objects reconstructs everyday items and complex scenes.
    • SAM 3D Body estimates human body shapes and poses with precision.
  • Benchmark Dataset: The SAM 3D Artist Objects dataset, built in collaboration with artists, sets a new standard for evaluating 3D reconstruction.
  • Applications Across Fields:
  • Robotics: Enhances spatial awareness for autonomous systems.
  • Sports Medicine: Supports advanced modeling for performance and rehabilitation.
  • AR/VR & Gaming: Accelerates asset creation for immersive worlds.
  • Commerce: Powers View in Room on Facebook Marketplace, letting buyers preview furniture or décor in their own spaces.

Segment Anything Playground

Meta’s Segment Anything Playground democratizes access to these models. Users can:

  • Upload images or videos and apply text prompts for segmentation.
  • Reconstruct 3D scenes and experiment with new perspectives.
  • Use templates for practical edits (pixelating faces, blurring license plates) or creative effects (motion trails, spotlight highlights).

Research & Open Source Commitment

  • SAM 3: Released with model weights, open vocabulary segmentation benchmarks, and annotation tools.
  • SAM 3D: Shared with checkpoints, inference code, and a novel benchmark dataset surpassing existing standards.
  • Meta’s Goal: To accelerate AI’s ability to understand and recreate the physical world, empowering both researchers and everyday creators.

Conclusion

With SAM 3 and SAM 3D, Meta is not just advancing AI — it’s democratizing creativity, enabling industries to reimagine workflows, and giving individuals tools to explore new dimensions of storytelling and utility. From healthcare to gaming, from commerce to art, these models signal a future where AI doesn’t just see the world — it helps us reshape it.

Reference: https://about.fb.com/news/2025/11/new-sam-models-detect-objects-create-3d-reconstructions/