Data Format of submission

Symbolic submissions: MIDI format with all sound events in program number 1 (solo piano). Track and channel are unrestricted.

Audio Submissions: wav format, 44.1kHz, 16-bit PCM


Training Datasets

Participants are welcome to train their systems on any dataset, including publicly available corpora, proprietary collections, or internally curated material. There are no restrictions on dataset origin, but we ask for full transparency.

Some suggested datasets for training and validation include:

  • ASAP[1]: A large dataset of classical piano performances sourced from MAESTRO, includes corresponding MIDI and audio. (n)ASAP[2] provides the alignment.
  • ATEPP[3]: A large dataset of transcribed MIDI expressive piano performances, organized by virtuosic performer. However, only around half of the dataset contains score MusicXML files.
  • VIENNA 4x22[4]: A small-scale dataset of 4 pieces with 22 different interpretations, including audio and MIDI and fine alignment.
  • Batik-plays-Mozart[5]: Fine-aligned performance MIDI dataset of Mozart played by Roland Batik.

Please clearly describe the datasets used for training and validation in your technical report. Important details to include are:

  • Dataset name or source
  • Size and number of pieces
  • Instrumentation and expressive characteristics
  • Data format (MIDI, audio, etc.)
  • Any preprocessing, cleaning, or augmentation steps applied

This helps the jury and the research community understand the representational capacity and limitations of each submission.


Post-Processing

To ensure fair evaluation, all post-processing applied to the preliminary round output must be documented in the submission report. Depending on your system type, please include the following:

  • Symbolic Output System: If your model generates symbolic MIDI output and you submit the sonified audio track, describe how audio is derived. Include soundfont names, software synths used (e.g., FluidSynth, Logic Pro), or player piano models.
    • If you would like to submit the MIDI output directly and allow us (the organizer team) for sonification, please contact huan.zhang@qmul.ac.uk during your submission.
  • Audio Output Systems: If your model outputs audio directly, describe if you have applied any enhancement steps such as EQ, reverb, compression, or noise reduction to the model’s output.
  • Controllability or Interventions: Clarify if the output is influenced by human-involved choices — such as selected tempo, dynamics range, segmentation, or annotated phrasing.
  • MIDI Cleanup: If symbolic outputs were manually edited (quantization, pedals, etc) before submission, that should be documented.

Submissions should aim for minimal human intervention. Manual correction is allowed only if it is well-documented and justified in the report.

References

  • [1] https://github.com/fosfrancesco/asap-dataset
  • [2] https://github.com/CPJKU/asap-dataset
  • [3] https://github.com/tangjjbetsy/ATEPP
  • [4] https://github.com/CPJKU/vienna4x22
  • [5] https://github.com/huispaty/batik_plays_mozart