GithubHelp home page GithubHelp logo

tangjjbetsy / atepp Goto Github PK

View Code? Open in Web Editor NEW
39.0 2.0 2.0 99.07 MB

ATEPP is a dataset of expressive piano performances by virtuoso pianists. (ISMIR2022)

Home Page: https://tangjjbetsy.github.io/ATEPP/

License: Creative Commons Zero v1.0 Universal

Shell 0.36% Python 99.64%
dataset midi music

atepp's Introduction

ATEPP: A Dataset of Automatically Transcribed Expressive Piano Performances

ATEPP is a dataset of expressive piano performances by virtuoso pianists. The dataset contains 11742 11674 performances (~1000 hours) by 49 pianists and covers 1580 1595 movements by 25 composers. All of the MIDI files in the dataset come from the piano transcription of existing audio recordings of piano performances. Scores in MusicXML format are also available for around half of the tracks. The dataset is organized and aligned by compositions and movements for comparative studies. More details are presented in the paper.

Downloade the ATEPP dataset

Please follow disclaimer.md to agree a disclaimer and download a latest version of ATEPP (~212MB).

Inference

You can inference your own track with the modified code and new checkpoint in piano_transcription-master. The env and setup are the same as https://github.com/bytedance/piano_transcription

python3 pytorch/inference.py --model_type=Regress_onset_offset_frame_velocity_CRNN --checkpoint_path=300000_iterations.pth --audio_path="resources/schumann_romanzen.mp3" --cuda

Released Versions

Version-1.2 (Latest Release!)

There are several issues found with the ATEPP Version-1.1:

  1. Corrupted Transcriptions (CTs) were found because of errors made by the transcription model when dealing with low-quality recordings (usually the live performances or old recordings). To detect the CTs, we compared the total note number and the duration of a performance with the medians for the same composition. We filtered out those which have similar duration to the midian but have much more/less notes with thresholds. We also mannually filtered out low-quality (annotated as low quality) audios by listening to them.

    • In total, 1264 audios were categorised as low-quality, having high possibility to lead to CTs. The corrupted refers to those confirmed corrupted with midi file.
    • In addition to low-quality audios, we annotated 1436 audios with background noise. These audios differ from low-quality audios in the way that they were transcribed with higher accuracy compared to those corresponding to low-quality audios. We suggest to filter out the corresponding midi files (tagged with low quality and background noise) when using the dataset.
    • As for live performances, we annotated applause if the recording contains that. We used high quality to refer to live recordings with good quality (clear, no applause, almost imperceptible background noise).
  2. Combined Movements (CMs) (one midi/audio consists of two or more movements) were found through a similar detection process of comparing the total note number and the duration with the medians. 7 were found, cut and relabeled.

  3. Error Labels (ELs) of composition were found when we manually verify the suspected pieces. 220 performances were found wrongly labelled and corrected.

The quality feature has been added to the metadata for clarifying the audio-related information. The repetition feature (not thoroughly analysed) refers to whether the performer plays the repetition sections or not.

Changed Statistics:

  • 11674 performances
  • 1595 movements

Version-1.1

!Updates: 65 Pieces Removed

When creating ATEPP version-1.0, we only applied movement-wise matching to remove erroneously downloaded audio. Now, we finished detecting repeated audios by audio-wise fingerprint matching. Only 65 audios were detected repeated, and the corresponding transcribed midi files were removed. The repeats.csv lists the repeated transcribed files that have been removed.

Changed Statistics:

  • 11677 performances
  • 1002 hours

Version-1.0

  • 11742 performances (in midi format)
  • 1007 hours
  • 1580 movements
  • 25 composers
  • 49 performers
  • 43% with scores

Related Works

We've released a Python package developed for linking classical music recording & track to the corresponding composition / movement, useful in cleaning up metadata in classical music datasets.

Package on PyPI: https://pypi.org/project/composition-entity-linker/

Contact

Citation

@inproceedings{zhang2022atepp,
  title={ATEPP: A Dataset of Automatically Transcribed Expressive Piano Performance},
  author={Zhang, Huan and Tang, Jingjing and Rafee, Syed Rifat Mahmud and Fazekas, Simon Dixon Gy{\"o}rgy},
  booktitle={ISMIR 2022 Hybrid Conference},
  year={2022}
}

License

CC BY 4.0

atepp's People

Contributors

tangjjbetsy avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar

Forkers

tbergman freddd13

atepp's Issues

Copyright concern of this dataset

Hi, can I get some detail of the copyright in this dataset? According to the papar, it said "we download each track from a corresponding open source audio at YouTube Music".

But I check the csv file that most Youtube music downloaded is provided by the Universal Music Group, and the performed pianists are generally not "died over 70 year" so I assuming their arrangements are under the copyright.

Would you detail the "open source audio" about how this open souce? And same concern for the sheet music. For the pieces in MuseScore, does there scores are non-provided licenses (so have copyright issue), or did provide the opensource licenses? MuseScore is open to public and non-careful management the copyright, so made things become sensitive.

This is a very promising dataset, but unfortunately I need to address the Univerisity's concerns regarding copyright before the further research.

Best regards <3

Google Colab

Hello, is there by chance a colab to infer with your model and checkpoint? Thank you so much

Inquiry

Hello, (I'm only a musician, not an expert in programming) but I got interested in the fact that note durations are more accurate in the midis you provide rather than in Bytedance's original system. I was wondering if you altered their method to achieve those results, if you want to comment on that a little! And I also can't wait to try your system, may I ask, will you release it soon? Thanks for the last response.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.