GithubHelp home page GithubHelp logo

Comments (5)

jrzaurin avatar jrzaurin commented on May 28, 2024

Hey @rruizdeaustri

I will have a more detail look, but in general here are some comments:

  1. Use a simpler model, forget about the wide component and use simply a deeptabular component with defaults. (review the code in your example since the optimizers and schedures are not correctly defined. The Trainer not throwing an error is intentional, I might change it, but just define your Trainer as
    trainer = Trainer(
        model,
        objective="binary",
        callbacks=[ModelCheckpoint(filepath="model_weights/wd_out")],
        metrics=[Accuracy],
    )
  1. The results with Transformer based models depend A LOT on the parameters, far more than in GBMs, where all, XGBoost, LightGBM and CatBoost perform almost to their best performance out of the box. You could have a look to this relatively old post see if it helps

I hope this helps and let me know how you get on with this, see if I can help more

from pytorch-widedeep.

rruizdeaustri avatar rruizdeaustri commented on May 28, 2024

Hi @jrzaurin,

I have made the modifications you suggested and results
make more sense now. I'm optimising hyper-parameters
with optima in resnet and transformer models but the results are
far from the one got with LightGMB: AUC ~0.93 versus ~ 0.98 for lgqbm

Thanks !

Rbt

from pytorch-widedeep.

jrzaurin avatar jrzaurin commented on May 28, 2024

Hey @rruizdeaustri

Thanks for sharing the results :)

0.05 is perhaps a bit too much, maybe I can look at some examples if you would be willing to share them. However, I am afraid that this is the "brutal" reality for most (true) real world cases when it comes to DL vs GBMs.

You could try some other libraries see if their implementations are better or you get better results (?)

In my experience I have used DL for tabular data in a few occasions, but never aimed to beat GBMs, since I knew was a lost battle.

from pytorch-widedeep.

rruizdeaustri avatar rruizdeaustri commented on May 28, 2024

Hi @jrzaurin,

Yes, these are too much differences !

I could share with you the files I'm using to train as well as the data if you like. Let me know !

Thanks !

from pytorch-widedeep.

jrzaurin avatar jrzaurin commented on May 28, 2024

Hey @rruizdeaustri !

I am traveling at the moment, but if you join the slack channel we can move the conversation there and we can share the files. See if I have the time to give it a go myself! :)

Thanks!

from pytorch-widedeep.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.