Comments (3)
Hi Linfeng, for these large datasets, we trained on A100 gpus with 80Gb memory. These datasets contain large number of variates, so 32GB seems not sufficient unless you reduce the number of input token.
from patchtst.
Hi, it depends on the prediction length, batch size, look back window.... But generally speaking for those two large datasets we often use 4 (could be up to 8) 3090 or A5000 GPUs. Also we decrease the batch size. For the other datasets, one 3090 would be sufficient.
from patchtst.
Thank you for your quick reply! However, I am specifically referring to the pre-training phase which does not have prediction length since we're reconstructing masked patches.
Also I noticed that the default context length for pre-training is 512, which is different from the look-back window length used in downstream forecasting tasks. Just would like to confirm if this is intended.
from patchtst.
Related Issues (20)
- how to use learner.distributed(), in self supervised pretrain code ?
- How does the visualization of Attention Weights organize the code? HOT 2
- scale
- RevIN and StandardScaler HOT 15
- Multivariate Time Series Classification HOT 6
- .
- Error during installation ”Could not find a version that satisfies the requirement numpy==1.21.3“,but the actual version is now 1.26.4
- Performance about self-supervised learning
- Can this model fit unbalanced panel data(more than one individual)?
- 请问论文中的注意力矩阵在代码里怎么输出
- Multivariate predict univariate HOT 2
- Stock Price Forecasting using PatchTST model HOT 4
- Pretrained Models in Huggingface Repository
- Question about not applying inverse_transform HOT 2
- Question about Table 9.
- outputs和batch_y 序列长度问题
- question about attention layer shared weight
- RuntimeError: required rank 4 tensor to use channels_last format
- How and where to save preds.npy and trues.npy after fine-tuning correctly?
- Why the train loss is higher than the validation and test?
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from patchtst.