Comments (2)
Hello @tojiboyevf, thank you for you interest,
Q: In optimizer_step() function in main.py file you multiply lr by lr_scale till 650 steps but you didn't mention this part in paper. Do you do the same warm up for Adam and AdamW?
A: Yes, we did utilize warm-up in our implementation. Although it did not significantly improve performance, it helped stabilize the models during the first epochs. We applied the warm-up strategy to SGD, Adam and AdamW optimizers.
Q: How do you select the learning rate for SGD, Adam, AdamW when you increase the batch size? For instance, some authors in self-supervised models select the learning rate by this formula: lr = base_lr * batch_size / 256 and base_lr can be 0.2, 0.3 and other values.
A: For SGD, we set the learning rate to 0.03 when the batch size is 100-120 places (which corresponds to 400-480 images). You can adjust the learning rate proportionally based on the batch size. In our case, you can use lr = 0.05 * BS / 120 for SGD and 0.0002 * BS / 120 for AdamW.
Q: Do you use the same scheduler with the same settings for Adam and AdamW optimizers?
A: Yes, we use the same scheduler for all optimizers. Specifically, we multiply the learning rate by 0.3 after every 5 epochs. Although we experimented with different strategies, we found that they yielded similar performance.
Q: Did you use any framework to find the best hyperparameters?
A: No, we just tried some values and went with those that performed best on pitts30-val.
from mixvpr.
Cool! Thanks for your answers!
from mixvpr.
Related Issues (20)
- why is the skip connection executed inside the FeatureMixerLayer here? HOT 2
- Really a good work with simple but effective approach!
- Training loss and generalization during test HOT 4
- Just for testing HOT 2
- A singleGPU will run the results, but multiple GPUs will make an error!! HOT 3
- can you provide more detailed comparative data? HOT 2
- question about backbone ‘Swin’ HOT 5
- Multi-similarity mining on Pittsburgh30k training set HOT 6
- how to change query image shape? HOT 1
- The error samples are due to issues with the ground truth annotations rather than errors in the model predictions. HOT 3
- Dataset HOT 4
- About the specific number of images of the Mapillary Challenge dataset
- would you release the resnet18 pretrained model?
- How to evaluate this model?
- Some questions
- About pitts30k_val.mat. I can not find it in PittsburgDataset HOT 1
- License File
- Releasing the model on torch.hub?
- On releasing the trained weights of ResNet50 with 2048 dimensionality.
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from mixvpr.