GithubHelp home page GithubHelp logo

vanint / datasetexpansion Goto Github PK

View Code? Open in Web Editor NEW
73.0 73.0 2.0 51.44 MB

The repository of Expanding Small-Scale Datasets with Guided Imagination (NeurIPS 2023).

Python 30.13% Shell 0.10% Jupyter Notebook 69.76%

datasetexpansion's People

Contributors

vanint avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar

datasetexpansion's Issues

Data augmentation

Hello seniors, can this data augmentation method be used in the field of speech.

environment

Hello, I would like to ask if the environment for this code has requirements for torch and python.Thanks

Dataset Expansion

Hello, I would like to confirm that the "Dataset Expansion" step is for training. It does not directly provide us with the checkpoint after training. Is that so?

MedMNIST dataset

Hi!

I am wondering how you save MedMNIST dataset on your local directory.
To my understanding, we can download .npz file for each dataset from the official website(https://zenodo.org/records/10519652).
Based on your implementation, it seems you saved them as .png or .jpg files in the different folders for each label.
Do you have the code to share?

dataset replication

Is it possible to perform data augmentation with a homemade railroad scene dataset, and if so, how should it be done? Thanks.

Details about finetuning SD on medical image

Hello.

I am currently trying your work on medical domain based on Dreambooth.

But there are a lot of missing details on finetuning with SD in this paper (Only number of training image mentioned).

Can you tell about more details based on the repository above? (e.g. class word, prompt, training time, etc.)

About enviroment configuration

Hi, thank you for your work. Initialize GIF_ Encountered some issues while modeling. It cannot your code run directly.
when I run the following:

model = instantiate_from_config(config.model)

I got the following error message:
/data/zhuhaowei/anaconda/anaconda3/envs/data-expand/lib/python3.10/site-packages/torchvision/transforms/transforms.py:329: UserWarning: Argument 'interpolation' of type int is deprecated since 0.13 and will be removed in 0.15. Please use InterpolationMode enum. warnings.warn( /data/zhuhaowei/anaconda/anaconda3/envs/data-expand/lib/python3.10/site-packages/pytorch_lightning/utilities/distributed.py:258: LightningDeprecationWarning: pytorch_lightning.utilities.distributed.rank_zero_onlyhas been deprecated in v1.8.1 and will be removed in v2.0.0. You can import it frompytorch_lightning.utilitiesinstead. rank_zero_deprecation( ==> Preparing dataset fgvc_aircraft Load fgvc_aircraft data finished. ==> creating model 'CLIP-VIT-B32' Model CLIP loaded. Global Step: 470000 LatentDiffusion: Running in eps-prediction mode DiffusionWrapper has 859.52 M params. Working with z of shape (1, 4, 32, 32) = 4096 dimensions. Traceback (most recent call last): File "/data/zhuhaowei/code/DatasetExpansion-main/GIF_SD/custom/dataset_expansion_stable_diffusion_CLIP_batch_optimization_final.py", line 732, in <module> main() File "/data/zhuhaowei/code/DatasetExpansion-main/GIF_SD/custom/dataset_expansion_stable_diffusion_CLIP_batch_optimization_final.py", line 555, in main GIF_model = load_model_from_config(config, f"{args.ckpt}") File "/data/zhuhaowei/code/DatasetExpansion-main/GIF_SD/custom/dataset_expansion_stable_diffusion_CLIP_batch_optimization_final.py", line 445, in load_model_from_config model = instantiate_from_config(config.model) File "/data/zhuhaowei/code/DatasetExpansion-main/GIF_SD/custom/ldm/util.py", line 85, in instantiate_from_config return get_obj_from_str(config["target"])(**config.get("params", dict())) File "/data/zhuhaowei/code/DatasetExpansion-main/GIF_SD/custom/ldm/models/diffusion/ddpm.py", line 461, in __init__ self.instantiate_cond_stage(cond_stage_config) File "/data/zhuhaowei/code/DatasetExpansion-main/GIF_SD/custom/ldm/models/diffusion/ddpm.py", line 519, in instantiate_cond_stage model = instantiate_from_config(config) File "/data/zhuhaowei/code/DatasetExpansion-main/GIF_SD/custom/ldm/util.py", line 85, in instantiate_from_config return get_obj_from_str(config["target"])(**config.get("params", dict())) File "/data/zhuhaowei/code/DatasetExpansion-main/GIF_SD/custom/ldm/modules/encoders/modules.py", line 141, in __init__ self.tokenizer = CLIPTokenizer.from_pretrained(version) File "/data/zhuhaowei/anaconda/anaconda3/envs/data-expand/lib/python3.10/site-packages/transformers/tokenization_utils_base.py", line 1784, in from_pretrained return cls._from_pretrained( File "/data/zhuhaowei/anaconda/anaconda3/envs/data-expand/lib/python3.10/site-packages/transformers/tokenization_utils_base.py", line 1825, in _from_pretrained init_kwargs = json.load(tokenizer_config_handle) File "/data/zhuhaowei/anaconda/anaconda3/envs/data-expand/lib/python3.10/json/__init__.py", line 293, in load return loads(fp.read(), File "/data/zhuhaowei/anaconda/anaconda3/envs/data-expand/lib/python3.10/json/__init__.py", line 346, in loads return _default_decoder.decode(s) File "/data/zhuhaowei/anaconda/anaconda3/envs/data-expand/lib/python3.10/json/decoder.py", line 337, in decode obj, end = self.raw_decode(s, idx=_w(s, 0).end()) File "/data/zhuhaowei/anaconda/anaconda3/envs/data-expand/lib/python3.10/json/decoder.py", line 355, in raw_decode raise JSONDecodeError("Expecting value", s, err.value) from None json.decoder.JSONDecodeError: Expecting value: line 1 column 1 (char 0)

I think there are some issues with my environment configuration. Now I am using transformers=4.19.2, torch=1.13.1, torch vision=0.14.1, and pytorch_ Lighting=1.9.0.

Can you provide a more detailed version of the environment? Thanks!

About training errors.

I trained on the cifar dataset, but encountered some problems. In dataset_expansion_stable_diffusion_CLIP_batch_optimization_final, bugs appear at this location in the program.
image
The model has problems in calculating the gradient, which hopefully can be solved later.
image

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.