The submission deadline is May 15th 2023 23:59 Anywhere on Earth (AoE)


Challenge submission consists of a submission package (one zip package) containing system outputs, system meta information, and technical report (pdf file).

Submission process shortly:

  1. Participants run their system with an evaluation dataset, and produce the system output in the specified format. Participants are allowed to submit 4 different system outputs per task or subtask.
  2. Participants create a meta-information file to go along the system output to describe the system used to produce this particular output. Meta information file has a predefined format to help the automatic handling of the challenge submissions. Information provided in the meta file will be later used to produce challenge results. Participants should fill in all meta information and make sure meta information file follows defined formatting.
  3. Participants describe their system in a technical report in sufficient detail. There will be a template provided for the technical report.
  4. Participants prepare the submission package (zip-file). The submission package contains system outputs, a maximum of 4 per task, systems meta information, and the technical report.
  5. Participants submit the submission package and the technical report to DCASE2023 Challenge.

Please read carefully the requirements for the files included in the submission package!

Submission system

The link to submission system will be added close to the submission deadline.

By submitting to the challenge, participants agree for the system output to be evaluated and to be published together with the results and the technical report on the DCASE Challenge website under CC-BY license.

Submission package

Participants are instructed to pack their system output(s), system meta information, and technical report into one zip-package. Example package will be added close to the submission deadline.

Submission label

A submission label is used to index all your submissions (systems per tasks). To avoid overlapping labels among all submitted systems, use the following way to form your label:

[Last name of corresponding author]_[Abbreviation of institute of the corresponding author]_task[task number][subtask letter (optional)]_[index number of your submission (1-4)]

For example, the baseline systems would have the following labels:

  • Martin_TAU_task1_1
  • Dohi_HIT_task2_1
  • Politis_TAU_task3_1
  • Turpault_INR_task4_1
  • Morfi_QMUL_task5_1
  • Gontier_INR_task6a_1
  • Xie_TAU_task6b_1

A script for checking the content of the submission package will be provided for selected tasks. In that case, please validate your submission package accordingly.

System outputs

Participants must submit the results for the provided evaluation datasets.

  • Follow the system output format specified in the task description.

  • Tasks are independent. You can participate in a single task or multiple tasks.

  • Multiple submissions for the same task are allowed (maximum 4 per task). Use a running index in the submission label, and give more detailed names for the submitted systems in the system meta information files. Please mark carefully the connection between the submitted systems and system parameters description in the technical report (for example by referring to the systems by using the submission label or system name given in the system meta information file).

  • Submitted system outputs will be published online on the DCASE2023 website later to allow future evaluations.

Meta information

In order to enable fast processing of the submissions and meta analysis of submitted systems, participants should provide meta information presented in a structured and correctly formatted YAML-file. Participants are advised to fill in the meta information carefully while making sure all asked information is correctly provided.

A complete meta file will help us notice possible errors before officially publishing the results (for example unexpectedly large difference in performance between development and evaluation set) and allow contacting the authors in case we consider it necessary. Please note that task organizers may ask you to update the meta file after the challenge submission deadline.

Example meta files will be added closer to the submission deadline.

Technical report

All participants are expected to submit a technical report about the submitted system, to help the DCASE community better understand how the algorithm works.

Technical reports are not peer-reviewed. The technical reports will be published on the challenge website together with all other information about the submitted system. For the technical report, it is not necessary to follow closely the scientific publication structure (for example there is no need for extensive literature review). The report should however contain a sufficient description of the system.

Please report the system performance using the provided cross-validation setup or development set, according to the task. For participants taking part in multiple tasks, one technical report covering all tasks is sufficient, if the systems have only small differences. Describe the task-specific parameters in the report.

Participants can also submit the same report as a scientific paper to DCASE 2023 Workshop. In this case, the paper must respect the structure of a scientific publication, and be prepared according to the provided Workshop paper instructions and template. Please note that the template is slightly different, and you will have to create a separate submission to the DCASE2023 Workshop track in the submission system. Please refer to the workshop webpage for more details. DCASE2023 Workshop papers will be peer-reviewed.


Reports are in format 4+1 pages. Papers are maximum 5 pages, including all text, figures, and references, with the 5th page containing only references.

The templates for technical report will be added later.