SAM produces high-quality object masks from input prompts and can be used to generate masks for all objects in an image.
The Segment Anything Model (SAM) is an object mask generation model that produces high-quality object masks from input prompts such as points or boxes, and it can be used to generate masks for all objects in an image. It has been trained on a dataset of 11 million images and 1.1 billion masks, and has strong zero-shot performance on a variety of segmentation tasks. The code requires Python 3.8 or higher, as well as PyTorch 1.7 and TorchVision 0.8, and can be installed via pip install or by cloning the repository locally. SAM's lightweight mask decoder can be exported to ONNX format and used in any environment that supports ONNX runtime. The code is licensed under Apache 2.0 and contributions are welcome.