GithubHelp home page GithubHelp logo

Comments (13)

SreeHarshaNelaturu avatar SreeHarshaNelaturu commented on May 10, 2024

REDS trained EDVR needs to be run through two stages, so you could try passing the output from stage 1 through the stage 2 model.
Other than that I believe there is a flip_test mode, which helps improve the quality.

from mmagic.

jorenvs avatar jorenvs commented on May 10, 2024

Yeah I figured out it has two stages afterwards, I'll rerun the experiment again soon.
The flip_test mode looks interesting, I guess an ensemble should increase generalisation a little.

from mmagic.

adamsvystun avatar adamsvystun commented on May 10, 2024

I can report a similar issue, I have tried using both stages, using flip_test, tested on a variety of videos. The model does not perform to the level it does on REDS4. It has multiple artifacts, and the quality overall is blurry.

from mmagic.

adamsvystun avatar adamsvystun commented on May 10, 2024

Okay, I solved my issue. The problem was the downsampling method. The datasets, that the model was trained on, were created by downsampling with MATLAB's imresize function. So if you generate input data with anything else (opencv, ffmpeg) - it doesn't work. You have to use MATLAB's imresize, or it's python's equivalent, which is implemented in this repo here.

from mmagic.

jorenvs avatar jorenvs commented on May 10, 2024

Hmm, that's what I feared. That kind of defeats the purpose of super resolution. I don't want to downsample my data, I want to upsample it :).

from mmagic.

SreeHarshaNelaturu avatar SreeHarshaNelaturu commented on May 10, 2024

@adamsvystun
Could you mention the exact flow that you used with the function you mentioned. did you basically send your input video through that method say if it H x W -> Target Resolution and subsequently pass it through EDVR to get the output. There were some weird blue-green artifacts during fast motion in my output, so I'm curious.

from mmagic.

adamsvystun avatar adamsvystun commented on May 10, 2024

@jorenvs It should work with upsampling. In my case I had a video in 720p and wanted to test 180p->720p upsampling, that's why I had to downsample. And it turns out that the model is very sensitive to the way you are doing this. If you have a video in low res only, it should just work.

@SreeHarshaNelaturu Yeah, for testing, I first downsample, then upsample with the model, and compare the results. Not sure about blue-green artifacts, I did not have any.

from mmagic.

jorenvs avatar jorenvs commented on May 10, 2024

well, my videos are 1344x1344, so not really low quality. That's all relative to the angle of the lens of course, these are generated from 360° 5.6k gopro videos. The goal is to be able read far away text on traffic signs and such.

from mmagic.

SreeHarshaNelaturu avatar SreeHarshaNelaturu commented on May 10, 2024

Thank you for the prompt response @adamsvystun I was wondering about the part you'd mentioned about not using FFmpeg or cv2 to generate input data. What did you use to extract frames from the video to SR in your case to using those methods.

from mmagic.

adamsvystun avatar adamsvystun commented on May 10, 2024

@SreeHarshaNelaturu I said don't use FFmpeg or cv2 for downscaling (resizing down). For frame extraction you can use anything you want.

from mmagic.

SreeHarshaNelaturu avatar SreeHarshaNelaturu commented on May 10, 2024

Gotcha, I think the blue-green error is a consequence of something else.
And yep, I was resizing via FFmpeg, might help to resize after extraction.

Thank you!

from mmagic.

ryul99 avatar ryul99 commented on May 10, 2024

I'm not sure but I think downscaling matlab's method could differ from FFmpeg and cv2
In my case, EDVR works well with the bicubic downscaling method but it has artifact like this with others (ex: low-res videos from youtube).
I guess that EDVR which trained with REDS dataset is overfitted about the reconstruction of bicubic downscaling as REDS dataset is consist of the bicubic downscaled dataset

from mmagic.

xinntao avatar xinntao commented on May 10, 2024

Yes, the current CNN-based methods does generalize to other datasets with different downsampling kernels.

There is another research filed called blind SR to solve this issue.

from mmagic.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.