--- tags: - image-to-3d - pytorch_model_hub_mixin - model_hub_mixin library_name: mast3r repo_url: https://github.com/naver/mast3r --- ## Grounding Image Matching in 3D with MASt3R ```bibtex @misc{mast3r_arxiv24, title={Grounding Image Matching in 3D with MASt3R}, author={Vincent Leroy and Yohann Cabon and Jerome Revaud}, year={2024}, eprint={2406.09756}, archivePrefix={arXiv}, primaryClass={cs.CV} } @inproceedings{dust3r_cvpr24, title={DUSt3R: Geometric 3D Vision Made Easy}, author={Shuzhe Wang and Vincent Leroy and Yohann Cabon and Boris Chidlovskii and Jerome Revaud}, booktitle = {CVPR}, year = {2024} } ``` # License The code is distributed under the CC BY-NC-SA 4.0 License. See [LICENSE](https://github.com/naver/mast3r/blob/main/LICENSE) for more information. For the checkpoints, make sure to agree to the license of all the public training datasets and base checkpoints we used, in addition to CC-BY-NC-SA 4.0. The mapfree dataset license in particular is very restrictive. For more information, check [CHECKPOINTS_NOTICE](https://github.com/naver/mast3r/blob/main/CHECKPOINTS_NOTICE). # Model info Gihub page: https://github.com/naver/mast3r/ | Modelname | Training resolutions | Head | Encoder | Decoder | |-------------|----------------------|------|---------|---------| | MASt3R_ViTLarge_BaseDecoder_512_catmlpdpt_nonmetric | 512x384, 512x336, 512x288, 512x256, 512x160 | CatMLP+DPT | ViT-L | ViT-B | # How to use First, [install mast3r](https://github.com/naver/mast3r?tab=readme-ov-file#installation). To load the model: ```python from mast3r.model import AsymmetricMASt3R import torch model = AsymmetricMASt3R.from_pretrained("naver/MASt3R_ViTLarge_BaseDecoder_512_catmlpdpt_nonmetric") device = torch.device("cuda" if torch.cuda.is_available() else "cpu") model.to(device) ```