Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add ViTMatte model #25040

Closed
2 tasks done
amyeroberts opened this issue Jul 24, 2023 · 1 comment
Closed
2 tasks done

Add ViTMatte model #25040

amyeroberts opened this issue Jul 24, 2023 · 1 comment

Comments

@amyeroberts
Copy link
Collaborator

Model description

ViTMatte is a recently released model for alpha matting on images i.e. background removal.

The model accepts an input image and trimap (manually labelled grayscale image outlining the rough border of the foreground object) and predicts the alpha mate for each pixel.

It introduces a series of small adaptations to the ViT architecture - selective global attention + window attention; adding convolutional blocks between transformers blocks - to reduce computational complexity and enhancing the high-frequency information passed through the network.

At the time of publishing, ViTMatte showed SOTA performance on Distinctions-646 and strong performance (> Mattformer) on Composition-1K.

Open source status

  • The model implementation is available
  • The model weights are available

Provide useful links for the implementation

Github: https://github.com/hustvl/ViTMatte
Paper: https://arxiv.org/pdf/2305.15272.pdf
Demo: https://colab.research.google.com/drive/1Dc2qoJueNZQyrTU19sIcrPyRDmvuMTF3?usp=sharing

@NielsRogge
Copy link
Contributor

This can be closed now thanks to #25843

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

Successfully merging a pull request may close this issue.

2 participants