GithubHelp home page GithubHelp logo

Comments (9)

LiJunnan1992 avatar LiJunnan1992 commented on May 16, 2024

Hi, I used the 40GB gpu.
Data reading was not a major speed bottleneck, but you could try to resize the image on hard drive because some images could have a high resolution.

from albef.

shoutOutYangJie avatar shoutOutYangJie commented on May 16, 2024

from albef.

LiJunnan1992 avatar LiJunnan1992 commented on May 16, 2024

Using 8 A100 GPUs, it takes 2-3 days with 4M images, and around 7-8 days with 14M images. You could make training faster by reducing the image resolution to 224 and increasing the batch size, the performance would be roughly the same. You can also try some other memory reduction techniques such as zero-redundancy optimizer.

from albef.

shoutOutYangJie avatar shoutOutYangJie commented on May 16, 2024

from albef.

LiJunnan1992 avatar LiJunnan1992 commented on May 16, 2024

Yes I think you could.
The pre-training dataset annotation (image paths and text) is released.

from albef.

sunanhe avatar sunanhe commented on May 16, 2024

Hi, do you get similar performance using 8 V100 GPUs?

can i use 8 32gb-v100  gpus to reproduce your training result?  by the way, the code of data preprocessing (filter some pairs) will be released?

------------------ Original ------------------ From: Junnan Li @.> Date: Wed,Dec 8,2021 6:05 PM To: salesforce/ALBEF @.> Cc: shoutOutYangJie @.>, Author @.> Subject: Re: [salesforce/ALBEF] what size of your A100 gpu's memory? (Issue#31) Using 8 A100 GPUs, it takes 2-3 days with 4M images, and around 7-8 days with 14M images. You could make training faster by reducing the image resolution to 224 and increasing the batch size, the performance would be roughly the same. You can also try some other memory reduction techniques such as zero-redundancy optimizer. — You are receiving this because you authored the thread. Reply to this email directly, view it on GitHub, or unsubscribe. Triage notifications on the go with GitHub Mobile for iOS or Android.

from albef.

shoutOutYangJie avatar shoutOutYangJie commented on May 16, 2024

from albef.

yangbang18 avatar yangbang18 commented on May 16, 2024

@shoutOutYangJie
Hi, it seems that you had reproduced the results with 8 V100 GPUs.

  1. Did you use the same configurations as in Pretrain.yaml?
  2. How many hours per epoch it took for the training?
  3. Have you tried to reduce the image resolution from 384 to 224?

Looking forward to your reply.

from albef.

shoutOutYangJie avatar shoutOutYangJie commented on May 16, 2024

from albef.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.