You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+4-4Lines changed: 4 additions & 4 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -185,7 +185,7 @@ Navigate (```cd```) to the root of the toolbox ```[YOUR_CIRTORCH_ROOT]```.
185
185
<details>
186
186
<summary><b>Training</b></summary><br/>
187
187
188
-
Our code can be used to fine-tune networks with whiteningadded as an FC layer after the pooling.
188
+
This toolbox can be used to fine-tune networks with end-to-end whitening, i.e., whitening added as an FC layer after the pooling and learned together with the convolutions.
189
189
To train such a setup you should run the following commands (the performance will be evaluated every 5 epochs on `roxford5k` and `rparis6k`):
@@ -219,7 +219,7 @@ Navigate (```cd```) to the root of the toolbox ```[YOUR_CIRTORCH_ROOT]```.
219
219
220
220
- Whitening FC layer is initialized in a supervised manner using our training data and off-the-shelf features.
221
221
- Whitening FC layer is precomputed for popular architectures and pooling methods, see [imageretrievalnet.py#L50](https://github.com/filipradenovic/cnnimageretrieval-pytorch/blob/474b1fe61ff0e8a6f076ef58f7334cf33d7a3773/cirtorch/networks/imageretrievalnet.py#L50) for the full list of precomputed FC layers.
222
-
- When whitening is added in the fine-tuning procedure, we notice that the performance is highest if the images are with a similar high-resolution at train and test time.
222
+
- When whitening is added in the fine-tuning procedure, the performance is highest if the images are with a similar high-resolution at train and test time.
223
223
- When whitening is added, the distribution of pairwise distances changes significantly, so roughly twice larger margin should be used for contrastive loss. In this scenario, triplet loss performs slightly better.
224
224
- Additional tunning of hyper-parameters can be performed to achieve higher performance or faster training. Note that, in this example, `--neg-num` and `--image-size` hyper-parameters are chosen such that the training can be performed on a single GPU with `16 GB` of memory.
225
225
@@ -228,9 +228,9 @@ Navigate (```cd```) to the root of the toolbox ```[YOUR_CIRTORCH_ROOT]```.
228
228
<details>
229
229
<summary><b>Testing our pretrained networks with whitening learned end-to-end</b></summary><br/>
230
230
231
-
We also provide our end-to-end pre-trained networks, trained both on `retrieval-SfM-120k (rSfM120k)` and [`Google Landmarks 2018 (GL18)`](https://www.kaggle.com/google/google-landmarks-dataset) train datasets.
231
+
Pretrained networks with whitening learned end-to-end are provided, trained both on `retrieval-SfM-120k (rSfM120k)` and [`Google Landmarks 2018 (GL18)`](https://www.kaggle.com/google/google-landmarks-dataset) train datasets.
232
232
Whitening is learned end-to-end during the network training, so there is no need to compute it as a post-processing step, although one can do that, as well.
233
-
For example, multi-scale evaluation of ResNet101 with GeM and whitening trained on `Google Landmarks 2018 (GL18)` dataset using high-resolution images and a triplet loss, is performed with the following script:
233
+
For example, multi-scale evaluation of ResNet101 with GeM and end-to-end whitening trained on `Google Landmarks 2018 (GL18)` dataset using high-resolution images and a triplet loss, is performed with the following script:
0 commit comments