GithubHelp home page GithubHelp logo

voice_conversion's People

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar

voice_conversion's Issues

The code in solver.py seems wrong.

Hi @jjery2243542

Thank you for your code. When I read the code, I find there is something wrong:

For master branch:

No 'train' mode in function 'solver.Solver.train', but in main.py, I see the code 'solver.train(args.output_model_path, args.flag, mode='train')'.

Please have a check if you have time.

length.txt file for data preprocessing.

Hi, Thank you for the code. I am not able to get the length.txt file you use to create the json files. Also the implementation you've shared doesn't seem to have a training routine for the stage 1 vae can you share the training details for stage one followed by stage two. It would be massively helpful if you can include step wise training procedure in your read me.

log_f0 does n't exist

Hello,

I was able to run make_dataset_vctk.py and get a h5py output. But the mcep normalize fails:
python mcep_normalize.py VCTK_h5py VCTK_h5py_norm
processing speaker_id=225
Traceback (most recent call last):
File "mcep_normalize.py", line 13, in
utt_f0 = [f_in[f'train/{speaker}/{utt_id}/log_f0'][()] for utt_id in f_in[f'train/{speaker}'].keys()]
File "mcep_normalize.py", line 13, in
utt_f0 = [f_in[f'train/{speaker}/{utt_id}/log_f0'][()] for utt_id in f_in[f'train/{speaker}'].keys()]
File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
File "/home/tfs/venv_voice_conversion_jjerry/lib/python3.6/site-packages/h5py/_hl/group.py", line 177, in getitem
oid = h5o.open(self.id, self._e(name), lapl=self._lapl)
File "h5py/_objects.pyx", line 54, in h5py._objects.with_phil.wrapper
File "h5py/_objects.pyx", line 55, in h5py._objects.with_phil.wrapper
File "h5py/h5o.pyx", line 190, in h5py.h5o.open
KeyError: "Unable to open object (object 'log_f0' doesn't exist)"

Apparently, you are saving mel and lin in the make_dataset*.py modules. but you are reading log_f0, mc_mean, mc_std, f0_mean, f0_std in your mcep_normalize as well as your convert.py codes.

Is your documentation missing a step that arrives at log_f0, mc_mean, mc_std, f0_mean, f0_std ?

Please clarify

Thanks,
Buvana

How to get the .h5 file ?

Dear Developer,
Where is the .h5 file whose path is defined inside the code ? Do we need to compute it some way or is it present with the code?

Torch not compiled with CUDA enabled

Hi. I think your code has a CUDA dependency. However, I do not have CUDA or any GPU available. When I load my own saved model, I get the error "Torch not compiled with CUDA enabled" . Can you please modify the inference.py code so that I could test my model?
Thanks for your help.

Saving the model

The main.py file has been running for the past 16 hours and I also specified the output model path but there is not any model that is saved in that path ?
Am I missing something ? How do I save the model?
I am getting this window
Screenshot from 2019-08-22 10-12-58

vctk.json file for pre-trained model?

I was trying to run the pre-trained model but there is a file whose description is unknown and that is vctk.json. Did you already provide this file or is it missing ?

Error in self.permute_data(data)

Hello,
I was trying to run the main.py file to retrain the model as you guided but in the middle of the process it threw an error "too many values to unpack (expected 2)" at line 279 in solver.py which is (c_i, c_j), (x_i_t, x_i_tk, x_i_prime, x_j) = self.permute_data(data).
Then I recognized that permute() function does not output many elements like this.

Did I miss something?
Can anybody help me out of it?

Running make_single_samples.py and the paths?

I am trying to run the make_single_samples.py and not sure what to pass in the

  1. n_samples
  2. segment length
  3. training sampled index path (.json)

Can you please help ?
print('usage: python3 make_single_samples.py [h5py path] [training sampled index path (.json)] '
'[n_samples] [segment length] [used speaker file path]')

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.