Reading @MetaAI's Segment-Anything, and I believe today is one of the "GPT-3 moments" in computer vision. It has learned the *general* concept of what an "object" is, even for unknown objects, unfamiliar scenes (e.g. underwater & cell microscopy), and ambiguous cases.
I still can't believe both the model and data (11M images, 1B masks) are OPEN-sourced. Wow.
What's the secret sauce? Just follow the foundation model mindset:
1. A very simple but scalable architecture that takes multimodal prompts: text, key points, bounding boxes.
2. Intuitive human annotation pipeline that goes hand-in-hand with the model design.
3. A data flywheel that allows the model to bootstrap itself to tons of unlabeled images.
IMHO, Segment-Anything has done everything right.
I still can't believe both the model and data (11M images, 1B masks) are OPEN-sourced. Wow.
What's the secret sauce? Just follow the foundation model mindset:
1. A very simple but scalable architecture that takes multimodal prompts: text, key points, bounding boxes.
2. Intuitive human annotation pipeline that goes hand-in-hand with the model design.
3. A data flywheel that allows the model to bootstrap itself to tons of unlabeled images.
IMHO, Segment-Anything has done everything right.
Introducing Segment Anything: Working toward the first foundation model for image segmentation
We're releasing the Segment Anything Model (SAM) — a step toward the first foundation model for image segmentation — and the SA-1B dataset.
ai.facebook.com
In addition to the new model, we’re also releasing the SA-1B dataset, which is 400x larger than any existing segmentation dataset — we hope this work will help accelerate computer vision research and enable entirely new applications.
<a href="GitHub - facebookresearch/segment-anything: The repository provides code for running inference with the SegmentAnything Model (SAM), links for downloading the trained model checkpoints, and example notebooks that show how to use the model.
Segment Anything
Meta AI Research, FAIRAlexander Kirillov, Eric Mintun, Nikhila Ravi, Hanzi Mao, Chloe Rolland, Laura Gustafson, Tete Xiao, Spencer Whitehead, Alex Berg, Wan-Yen Lo, Piotr Dollar, Ross Girshick
[Paper] [Project] [Demo] [Dataset]
The Segment Anything Model (SAM) produces high quality object masks from input prompts such as points or boxes, and it can be used to generate masks for all objects in an image. It has been trained on a dataset of 11 million images and 1.1 billion masks, and has strong zero-shot performance on a variety of segmentation tasks.
Segment Anything | Meta AI"
demo site:
Last edited: