Effortless AI-assisted data labeling with AI support from YOLO, Segment Anything (SAM+SAM2), MobileSAM!!
-
Updated
Mar 14, 2025 - Python
Effortless AI-assisted data labeling with AI support from YOLO, Segment Anything (SAM+SAM2), MobileSAM!!
Labeling tool with SAM(segment anything model),supports SAM, SAM2, sam-hq, MobileSAM EdgeSAM etc.交互式半自动图像标注工具
Tailor是一款视频智能裁剪、视频生成和视频优化的视频剪辑工具。目前的目标是通过人工智能技术减少视频剪辑的繁琐操作,让普通人也能简单实现专业剪辑人的水准!长远目标是让视频剪辑实现真正的AIGC!
Video-Inpaint-Anything: This is the inference code for our paper CoCoCo: Improving Text-Guided Video Inpainting for Better Consistency, Controllability and Compatibility.
[CVPR 2025] The code for "VideoRefer Suite: Advancing Spatial-Temporal Object Understanding with Video LLM"
Playground Web UI using segment-anything-2 models from the Meta.
Ultralytics VSCode snippets plugin to provide quick examples and templates of boilerplate code to accelerate your code development and learning.
A gradio based webui for meta segment-anything-model 2 (SAM2), both image and video are supported
Simple Video Summarization using Text-to-Segment Anything (Florence2 + SAM2) This project provides a video processing tool that utilizes advanced AI models, specifically Florence2 and SAM2, to detect and segment specific objects or activities in a video based on textual descriptions.
Image segmentation application that utilizes the SAM2 (Segment Anything Model) via API to perform object detection and segmentation on uploaded images.
Use natural language to ground relevant things.
Run Segment Anything 2 (SAM 2) on macOS using Core ML models
A script that utilises Facebook's SAM-2 model to add segmentation mask, bounding box, and rotation angle annotations to the MIDV500 and MIDV2019 datasets.
This repository provides a powerful AI-driven solution for removing objects from videos using text prompts. By integrating SAM2, Florence2, and ProPainter, the model enables precise and seamless object removal. Simply describe the objects to remove (e.g., "man, car, cap, basket"), and the AI will handle the rest with high accuracy.
A computer vision project in Python focused on open-image classification without the limitations of predefined datasets like COCO. It leverages the Segment Anything Model 2 (SAM2) for segmentation, a Large Vision-Language Model (LLaVA) for image description, and a Large Language Model (DeepSeek) for instance classification.
Official PyTorch implementation for class-wise segmentation of construction and demolition waste in cluttered environments
Add a description, image, and links to the sam2 topic page so that developers can more easily learn about it.
To associate your repository with the sam2 topic, visit your repo's landing page and select "manage topics."