g-astruc commited on
Commit
689ccbb
•
1 Parent(s): f8906fa

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -13
README.md CHANGED
@@ -6,27 +6,19 @@ license: mit
6
 
7
  [Guillaume Astruc](https://gastruc.github.io/), [Nicolas Gonthier](https://ngonthier.github.io/), [Clement Mallet](https://www.umr-lastig.fr/clement-mallet/), [Loic Landrieu](https://loiclandrieu.com/)
8
 
9
-
10
- Official models for [_AnySat: An Earth Observation Model for Any Resolutions, Scales, and Modalities_](https://arxiv.org/pdf/2404.08351.pdf)
11
-
12
-
13
  <p align="center">
14
  <img src="https://cdn-uploads.huggingface.co/production/uploads/662b7fba68ed7bbf40bfb0df/Jh9eOnMePFiL84TOzhe86.png" alt="image/png" width="600" height="300">
15
  </p>
16
 
17
  ## Abstract
18
 
19
- <div style="display: flex; align-items: center;">
20
- <div style="flex: 1;">
21
- <p>We introduce AnySat: a JEPA-based multimodal Earth Observation model that train simultaneously on diverse datasets with different scales, resolutions (spatial, spectral, temporal), and modality combinations.
22
 
 
23
 
24
- For more details and results, please check out our [github](https://github.com/gastruc/AnySat) and [project page](https://gastruc.github.io/projects/omnisat.html).</p>
25
- </div>
26
- <div style="flex: 1; display: flex; justify-content: center;">
27
- <img src="https://cdn-uploads.huggingface.co/production/uploads/662b7fba68ed7bbf40bfb0df/2tc0cFdOF2V0_KgptA-qV.png" alt="image/png" width="400"/>
28
- </div>
29
- </div>
30
 
31
  ### Inference 🔥
32
 
@@ -75,6 +67,11 @@ features = AnySat(data, scale=scale, keep_subpatch=True, modality_keep=modality)
75
 
76
  Note that the features will be of size 2*D. If you have several modalities of the same desired resolution, you should pick the most informative one (or modify the code to concatenate also the other modalities)
77
 
 
 
 
 
 
78
  To reproduce results, add new modalities, or do more experiments see the full code on [github]('https://github.com/gastruc/AnySat').
79
 
80
  ### Citing 💫
 
6
 
7
  [Guillaume Astruc](https://gastruc.github.io/), [Nicolas Gonthier](https://ngonthier.github.io/), [Clement Mallet](https://www.umr-lastig.fr/clement-mallet/), [Loic Landrieu](https://loiclandrieu.com/)
8
 
 
 
 
 
9
  <p align="center">
10
  <img src="https://cdn-uploads.huggingface.co/production/uploads/662b7fba68ed7bbf40bfb0df/Jh9eOnMePFiL84TOzhe86.png" alt="image/png" width="600" height="300">
11
  </p>
12
 
13
  ## Abstract
14
 
15
+ We introduce AnySat: a JEPA-based multimodal Earth Observation model that train simultaneously on diverse datasets with different scales, resolutions (spatial, spectral, temporal), and modality combinations.
 
 
16
 
17
+ For more details and results, please check out our [github](https://github.com/gastruc/AnySat) and [project page](https://gastruc.github.io/projects/omnisat.html).
18
 
19
+ <p align="center">
20
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/662b7fba68ed7bbf40bfb0df/2tc0cFdOF2V0_KgptA-qV.png" alt="image/png" width="400" height="200">
21
+ </p>
 
 
 
22
 
23
  ### Inference 🔥
24
 
 
67
 
68
  Note that the features will be of size 2*D. If you have several modalities of the same desired resolution, you should pick the most informative one (or modify the code to concatenate also the other modalities)
69
 
70
+
71
+ Example of use of AnySat:
72
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/662b7fba68ed7bbf40bfb0df/_x2ng-3c0jvLIP3R5WEwA.png)
73
+
74
+
75
  To reproduce results, add new modalities, or do more experiments see the full code on [github]('https://github.com/gastruc/AnySat').
76
 
77
  ### Citing 💫