Skip to content

pq-yang/MatAnyone2

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 

History

11 Commits
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation

MatAnyone Logo

Scaling Video Matting via a Learned Quality Evaluator

Peiqing Yang1โ€ƒ Shangchen Zhou1โ€ โ€ƒ Kai Hao1โ€ƒ Qingyi Tao2โ€ƒ
1S-Lab, Nanyang Technological Universityโ€ƒ 2SenseTime Research, Singaporeโ€ƒ
โ€ Project lead

MatAnyone 2 is a practical human video matting framework that preserves fine details by avoiding segmentation-like boundaries, while also shows enhanced robustness under challenging real-world conditions.

๐ŸŽฅ For more visual results, go checkout our project page


๐Ÿ“ฎ Update

  • [2026.03] Add uv, CLI, and huggingface support for easy installation and usage.
  • [2026.03] Release inference codes, evaluation codes, and gradio demo.
  • [2025.12] This repo is created.

๐Ÿ„๐Ÿปโ€โ™€๏ธ TODO

  • Release inference codes and gradio demo.
  • Release evaluation codes.
  • Release training codes for video matting model.
  • Release checkpoint and training codes for quality evaluator model.
  • Release real-world video matting dataset VMReal.

๐Ÿ”Ž Overview

overall_structure

๐Ÿ”ง Installation

Conda

  1. Clone Repo

    git clone https://github.com/pq-yang/MatAnyone2 cd MatAnyone2
  2. Create Conda Environment and Install Dependencies

    # create new conda env conda create -n matanyone2 python=3.10 -y conda activate matanyone2 # install python dependencies pip install -e . # [optional] install python dependencies for gradio demo pip3 install -r hugging_face/requirements.txt

uv

You may also install via uv:

# create a new project and add matanyone2 uv init my-matting-project && cd my-matting-project uv add matanyone2@git+https://github.com/pq-yang/MatAnyone2.git

๐Ÿ”ฅ Inference

Download Model

Download our pretrained model from MatAnyone 2 to the pretrained_models folder (pretrained model can also be automatically downloaded during the first inference).

The directory structure will be arranged as:

pretrained_models |- matanyone2.pth 

Quick Test

We provide some examples in the inputs folder. For each run, we take a video and its first-frame segmenatation mask as input. The segmenation mask could be obtained from interactive segmentation models such as SAM2 demo. For example, the directory structure can be arranged as:

inputs |- video |- test-sample1 # folder containing all frames |- test-sample2.mp4 # .mp4, .mov, .avi |- mask |- test-sample1.png # mask for targer person(s) |- test-sample2.png 

Run the following command to try it out:

# intput format: video folder python inference_matanyone2.py -i inputs/video/test-sample1 -m inputs/mask/test-sample1.png # intput format: mp4 python inference_matanyone2.py -i inputs/video/test-sample2.mp4 -m inputs/mask/test-sample2.png
  • The results will be saved in the results folder, including the foreground output video and the alpha output video.
  • If you want to save the results as per-frame images, you can set --save-image.
  • If you want to set a limit for the maximum input resolution, you can set --max-size, and the video will be downsampled if min(w, h) exceeds. By default, we don't set the limit.

Or you may directly run via CLI command:

matanyone2 -i inputs/video/test-sample1 -m inputs/mask/test-sample1.png
  • Run matanyone2 --help for a full list of options.

Python API ๐Ÿค—

You can load the model directly from Hugging Face using from_pretrained and run inference programmatically:

from matanyone2 import MatAnyone2, InferenceCore model = MatAnyone2.from_pretrained("PeiqingYang/MatAnyone2") processor = InferenceCore(model, device="cuda:0") processor.process_video( input_path="inputs/video/test-sample2.mp4", mask_path="inputs/mask/test-sample2.png", output_path="results", )

๐ŸŽช Interactive Demo

To get rid of the preparation for first-frame segmentation mask, we prepare a gradio demo on hugging face and could also launch locally. Just drop your video/image, assign the target masks with a few clicks, and get the the matting results!

We integrate MatAnyone Series in the demo. MatAnyone 2 is the default model. You can also choose MatAnyone as your processing model in "Model Selection".

cd hugging_face # install GUI dependencies pip3 install -r requirements.txt # FFmpeg required # launch the demo python app.py

By launching, an interactive interface will appear as follow.

overall_teaser

๐Ÿ“Š Evaluation

Please refer to the evaluation documentation for details.

๐Ÿ› ๏ธ Data Pipeline

data_pipeline

๐Ÿ“‘ Citation

If you find our repo useful for your research, please consider citing our paper:

@InProceedings{yang2026matanyone2, title = {{MatAnyone 2}: Scaling Video Matting via a Learned Quality Evaluator}, author = {Yang, Peiqing and Zhou, Shangchen and Hao, Kai and Tao, Qingyi}, booktitle = {CVPR}, year = {2026} } @inProceedings{yang2025matanyone, title = {{MatAnyone}: Stable Video Matting with Consistent Memory Propagation}, author = {Yang, Peiqing and Zhou, Shangchen and Zhao, Jixin and Tao, Qingyi and Loy, Chen Change}, booktitle = {CVPR}, year = {2025} }

๐Ÿ“ License

This project is licensed under NTU S-Lab License 1.0. Redistribution and use should follow this license.

๐Ÿ‘ Acknowledgement

This project is built upon MatAnyone and Cutie, with matting dataset files adapted from RVM. The interactive demo is adapted from ProPainter, leveraging segmentation capabilities from Segment Anything Model and Segment Anything Model 2. Thanks for their awesome works!

๐Ÿ“ง Contact

If you have any questions, please feel free to reach us at peiqingyang99@outlook.com.

About

[CVPR 2026] MatAnyone 2: Scaling Video Matting via a Learned Quality Evaluator

Resources

License

Stars

Watchers

Forks

Packages

 
 
 

Contributors