Edit model card

Grounding Image Matching in 3D with MASt3R

@misc{mast3r_arxiv24,
      title={Grounding Image Matching in 3D with MASt3R}, 
      author={Vincent Leroy and Yohann Cabon and Jerome Revaud},
      year={2024},
      eprint={2406.09756},
      archivePrefix={arXiv},
      primaryClass={cs.CV}
}

@inproceedings{dust3r_cvpr24,
      title={DUSt3R: Geometric 3D Vision Made Easy}, 
      author={Shuzhe Wang and Vincent Leroy and Yohann Cabon and Boris Chidlovskii and Jerome Revaud},
      booktitle = {CVPR},
      year = {2024}
}

License

The code is distributed under the CC BY-NC-SA 4.0 License. See LICENSE for more information.
For the checkpoints, make sure to agree to the license of all the public training datasets and base checkpoints we used, in addition to CC-BY-NC-SA 4.0.
The mapfree dataset license in particular is very restrictive. For more information, check CHECKPOINTS_NOTICE.

Model info

Gihub page: https://github.com/naver/mast3r/

Modelname Training resolutions Head Encoder Decoder
MASt3R_ViTLarge_BaseDecoder_512_catmlpdpt_nonmetric 512x384, 512x336, 512x288, 512x256, 512x160 CatMLP+DPT ViT-L ViT-B

How to use

First, install mast3r. To load the model:

from mast3r.model import AsymmetricMASt3R
import torch

model = AsymmetricMASt3R.from_pretrained("naver/MASt3R_ViTLarge_BaseDecoder_512_catmlpdpt_nonmetric")

device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
model.to(device)
Downloads last month
110,823
Safetensors
Model size
689M params
Tensor type
F32
Β·
Inference API
Inference API (serverless) does not yet support mast3r models for this pipeline type.

Spaces using naver/MASt3R_ViTLarge_BaseDecoder_512_catmlpdpt_metric 5

Collection including naver/MASt3R_ViTLarge_BaseDecoder_512_catmlpdpt_metric