Update README.md
Browse files
README.md
CHANGED
@@ -168,8 +168,8 @@ rectification, and grid interpolation. The methodology employed in each step is
|
|
168 |
|
169 |
- Inference speed: The ConvSwin2SR model demonstrates a commendable inference speed, particularly when handling a substantial batch of samples.
|
170 |
Specifically, when tasked with downscaling 248 samples, which is synonymous with processing data for an entire month at 3-hour intervals,
|
171 |
-
the model completes the operation in a mere 21 seconds. This level of efficiency is observed in a local computing environment outfitted with 16GB
|
172 |
-
|
173 |
|
174 |
# Evaluation
|
175 |
|
@@ -233,17 +233,23 @@ The Swin2 transformer optimizes its parameters using a composite loss function t
|
|
233 |
accuracy across different resolutions and representations:
|
234 |
|
235 |
1. **Primary Predictions Loss**:
|
236 |
-
- This term computes the L1 loss between the primary model predictions and the reference values. It ensures that the transformer's
|
237 |
-
closely match the ground truth
|
238 |
|
239 |
2. **Downsampled Predictions Loss**:
|
240 |
-
-
|
241 |
-
|
242 |
-
|
|
|
|
|
243 |
|
244 |
3. **Blurred Predictions Loss**:
|
245 |
- To ensure the model's robustness against small perturbations and noise, this term evaluates the L1 loss between blurred versions of the
|
246 |
-
predictions and the references. This encourages the model to produce predictions that maintain accuracy even under slight modifications
|
|
|
|
|
|
|
|
|
247 |
|
248 |
## Computing Infrastructure
|
249 |
|
|
|
168 |
|
169 |
- Inference speed: The ConvSwin2SR model demonstrates a commendable inference speed, particularly when handling a substantial batch of samples.
|
170 |
Specifically, when tasked with downscaling 248 samples, which is synonymous with processing data for an entire month at 3-hour intervals,
|
171 |
+
the model completes the operation in a mere 21 seconds. This level of efficiency is observed in a local computing environment outfitted with 16GB of
|
172 |
+
RAM and 4GB of GPU memory.
|
173 |
|
174 |
# Evaluation
|
175 |
|
|
|
233 |
accuracy across different resolutions and representations:
|
234 |
|
235 |
1. **Primary Predictions Loss**:
|
236 |
+
- This term computes the L1 loss between the primary model predictions and the reference values. It ensures that the transformer's
|
237 |
+
outputs closely match the ground truth.
|
238 |
|
239 |
2. **Downsampled Predictions Loss**:
|
240 |
+
- This term calculates the L1 loss between the downsampled versions of the predictions and the reference values. By incorporating this term,
|
241 |
+
the model is incentivized to preserve the underlying relations between both spatial resolutions. The references and predictions are upscaled
|
242 |
+
by average pooling by a factor of x5 to match the source resolution. Although this loss term could be (technically) computed with respect
|
243 |
+
to the low-resolution sample, the upscaled reference values are considered, due to the fact that the average pooling used for upscaling does
|
244 |
+
not represent the true relationship between both datasets considered.
|
245 |
|
246 |
3. **Blurred Predictions Loss**:
|
247 |
- To ensure the model's robustness against small perturbations and noise, this term evaluates the L1 loss between blurred versions of the
|
248 |
+
predictions and the references. This encourages the model to produce predictions that maintain accuracy even under slight modifications
|
249 |
+
in the data representation. On the other hand, it can smooth the prediction field too much, so it is a term whose use should be studied
|
250 |
+
before including it in your model. To produce the blurred values, a gaussian kernel of size 5 is applied.
|
251 |
+
|
252 |
+
By combining these loss terms, the ConvSwin2SR is trained to produce realistic predictions.
|
253 |
|
254 |
## Computing Infrastructure
|
255 |
|