Comments (8)
Thank you for your interest in my work. I'm glad to hear that you found it helpful.
Regarding the time taken to run the model, I did not pay close attention to the exact time but rather let it run for approximately 24 hours. While the model did overfit in this case, this is less of an issue in this particular domain. Based on my experience, I would expect that running the model for 500-1000 epochs would take 5-10 hours and would still yield useful results.
How come you cannot download the test dataset? I believe it is located at: https://reshare.ukdataservice.ac.uk/851861/?
Please let me know if there is anything else I can help you with!
from anoddpm.
Thank you very much for your reply, your reply is very helpful. I succeeded in training later, and it takes about a day. A neuroimaging dataset of brain tube patients (your test datasets) cannot be downloaded indeed, and the error shown in the figure below occurs.
And I still have some confusion about the work:
The train dataset: did you randomly extract a slice from each case 3D data and use it for training? That is, 100 patients correspond to 100 2D slices. The training set consists of 100 images in total.
The test dataset: did you sample 4 images per patient by np.linspace?
Looking forward to your reply.
from anoddpm.
I'm sorry but I cannot redistribute the dataset based on its terms of service. I also no longer have access to the original website as I don't have my old university logon. I would suggest reaching out to one of the authors from that data collection for permission.
So for training (from memory), I sampled one out of the 100 training volumes and then for that volume, I sampled a slice of the brain from that (the specific slice range is In the dataset code).
The test dataset, yes you are correct; I sampled 4 evenly spaced slices from the volume that contained the tumour using np.linspace. I looked through each volume and marked down the slice range of each tumour so that I could do that - the dictionary for which is located in that same dataset code
Hope this helps - let me know if there is anything further I can help with too
from anoddpm.
Thank you for your reply. My confusion has been answered. Wish everything goes well with your work.
from anoddpm.
Excuse me, what is the size of the data set for training your model (the total amount after data enhancement)
from anoddpm.
Thank you very much for your reply, your reply is very helpful. I succeeded in training later, and it takes about a day. A neuroimaging dataset of brain tube patients (your test datasets) cannot be downloaded indeed, and the error shown in the figure below occurs.
And I still have some confusion about the work: The train dataset: did you randomly extract a slice from each case 3D data and use it for training? That is, 100 patients correspond to 100 2D slices. The training set consists of 100 images in total. The test dataset: did you sample 4 images per patient by np.linspace?
Looking forward to your reply.
Hi, I also face the same problem that the test brain tube patients datasets cannot be downloaded, do you solve this problem? Thanks~
from anoddpm.
Excuse me, what is the size of the data set for training your model (the total amount after data enhancement)
Apologies I didn't get to this sooner. The approximate size is listed in my paper - it should be in the region of 100 volumes of 256x256x60 (60 relevant brain slices, the rest are discarded) - with additional up to +-3 degree's rotation and random translation of 0.02×width and 0.09×height.
As some of these are continuous ranges, it wouldn't be entirely possible to give a clear, accurate discrete count of the number of training images.
from anoddpm.
Thank you very much for your reply, your reply is very helpful. I succeeded in training later, and it takes about a day. A neuroimaging dataset of brain tube patients (your test datasets) cannot be downloaded indeed, and the error shown in the figure below occurs.
And I still have some confusion about the work: The train dataset: did you randomly extract a slice from each case 3D data and use it for training? That is, 100 patients correspond to 100 2D slices. The training set consists of 100 images in total. The test dataset: did you sample 4 images per patient by np.linspace?
Looking forward to your reply.Hi, I also face the same problem that the test brain tube patients datasets cannot be downloaded, do you solve this problem? Thanks~
Sadly I can't distribute the data in this case - I would recommend considering a different open source labelled anomalous dataset.
from anoddpm.
Related Issues (20)
- some doubts about the principle of the diffusion model HOT 1
- some quesion about detection HOT 5
- some questions about DDPM HOT 1
- Some problems with running detection HOT 2
- Batch Size HOT 3
- Where is the file of args used in the paper? HOT 2
- How Train and Test is split. HOT 2
- comparasion_models HOT 5
- What is a directory raw_cleaned ? HOT 7
- about the training time cost HOT 1
- Some problems about data HOT 1
- Encountered Access Issue of E-mail HOT 1
- Package Versions / Install Requirements? HOT 1
- how to get testData HOT 2
- high CPU usage discovered in training stage HOT 1
- result for MVTec HOT 1
- Asking about "import Comparative_models.CE as CE" HOT 2
- Problem about saving .mp4 files during training process HOT 3
- Access to training args.json used in the paper
- Pre-trained data HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from anoddpm.