Comments (3)
No. There is no particular reason.
When I wrote the paper, I cited mainly frameworks that were tested with the same evaluation metric in four benchmarks.
CESTa is also considered a good paper, but it was published in 2020. We cite more recent papers.
from compm.
No offense, I just find it hard to reach cesta's micro performance of 63.12 on dailydialog, which is 3 points higher than compm, one of the most advanced models nowadays, so I would like to ask if there is a reason why this is not comparable due to different experimental settings or different metrics settings.
I was wondering if perhaps they only considered micro as a performance criterion and your experiments took into account both micro and macro scores, and if this could be the basis for not using the data reported in their paper for comparison.
from compm.
CESTa was measured only with micro metrics in dailyDialog, and was not tested in EmoryNLP.
However, since CESTa shows good results in dailydialog, it is reasonable to select it as a comparative paper.
---caution---
I'm not fully aware of CESTa , but from a quick glance it seems that future utterances are taken into account as input. (In Introduction, Figure 2 seems to use the future feature.)
That is, if future utterances are considered input, they are not suitable as comparative papers. There are often studies that improve emotion recognition performance by using future utterances.
This is because typical ERC studies only use utterances before the current turn as context.
from compm.
Related Issues (13)
- Possible typo in the paper HOT 1
- I wonder if the model works fine when batch is not 1? HOT 5
- Google Drive link not found HOT 3
- "MELD_models/roberta-large/pretrained/no_freeze/emotion/1.0/model.bin" . It doesn't have mobel.bin file. HOT 2
- reproduce your work on MELD dataset, the weight F1 score is 66.52 HOT 2
- UnboundLocalError: local variable 'DATA_loader' referenced before assignment HOT 1
- RoBERTa large max input token size issue HOT 2
- Unexpected key(s) in state_dict: "context_model.embeddings.position_ids", "speaker_model.embeddings.position_ids" HOT 3
- Can I use this work for sentiment analysis of dyadic conversations? HOT 1
- pytorch-cuda version for the source code? HOT 2
- Can I ignore this warning? UndefinedMetricWarning: Precision and F-score are ill-defined and being set to 0.0 in labels with no predicted samples. HOT 1
- Why don't put batch_speaker_token on cuda? HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from compm.