ACM Multimedia 2026 Grand Challenge

AdoDAS: A Privacy-Preserving Multimodal Challenge for Adolescent Depression, Anxiety, and Stress Assessment

AdoDAS challenge supports multimodal modeling under strict privacy constraints by releasing representations and temporal metadata instead of raw recordings.

6,000
Participants
24,000
Audio-Video Segments
2
Challenge Tracks
7/1/2
Subject-Disjoint Split

Calling for Participations

We present AdoDAS, a privacy-preserving multimodal Grand Challenge dataset for adolescent Depression, Anxiety, and Stress (D/A/S) assessment. The benchmark combines a standardized reading passage with open-ended interview prompts and provides labels from DASS-21, including both subscale scores and item-level responses.

Adolescent mental health screening is critical for early identification and timely intervention, yet scalable assessment remains challenging in real-world practice. Depression, anxiety, and stress (D/A/S) are prevalent during adolescence and frequently co-occur, while routine screening still relies heavily on questionnaires and limited clinical resources. Recent progress in multimodal affective computing suggests that audio-visual behavioural cues during communication can provide signals complementary to self-reports, motivating standardised, reproducible benchmarks for multimodal D/A/S screening and assessment.

Despite growing interest, reproducible multimodal benchmarks for minors remain scarce. A key barrier is privacy and ethics: releasing raw audio/video recordings of children and adolescents can enable re-identification even after conventional de-identification, making open data sharing difficult. In addition, existing mental-health resources often differ in task definitions, data splits, and evaluation protocols, which hinders fair comparison and may introduce subject leakage.

We introduce AdoDAS, a feature-centric, privacy-preserving multimodal Grand Challenge for adolescent D/A/S assessment with labels derived from DASS-21. Data are collected using a tightly controlled school-based protocol that combines a standardised reading passage and open-ended question answering, balancing controlled elicitation and spontaneous expression while improving comparability across participants and sites. To enable reproducible research without exposing raw signals, AdoDAS withholds raw audio/video and releases pre-computed acoustic descriptors, visual representations, and cross-modal temporal metadatato support temporal modelling and multimodal fusion. In total, AdoDAS includes 6,000 participants and 24,000 audio-video segments across four sessions per participant.

The benchmark includes two tracks: (A1) multi-task binary classification for D/A/S screening and (A2) DASS-21 item-level prediction for interpretable symptom modeling. Both tracks use the same privacy-preserving feature inputs and official subject-disjoint split protocol, with ranking generated by the official script for deterministic evaluation.

Dataset

Participants and Ethics

All procedures follow institutional and school requirements for research involving minors. Written informed consent is collected from legal guardians and assent is obtained from participants. The benchmark release is designed to minimize re-identification risks.

Recording Protocol

Each participant contributes one scripted reading session (A01) and three open-ended sessions (B01-B03). Sessions are recorded under standardized conditions in quiet classrooms with fixed mouth-to-microphone distance and are capped at 60 seconds each.

Session Type Prompt
A01 Fixed Text Reading "The North Wind and the Sun" standardized reading passage.
B01 Open-ended Question Please describe how your day went yesterday.
B02 Open-ended Question Please describe your happiest memory from the past week.
B03 Open-ended Question Please describe your saddest memory from the past week.

Annotation and Labels

Ground truth is derived from DASS-21. For Track A1, each of D/A/S uses binary mapping: Normal (0) versus Mild-or-above (1). For Track A2, predictions are made at the 21-item response level (0-3), enabling interpretable symptom modeling.

Optional Auxiliary Attributes

Auxiliary demographic/context attributes are provided for stratified analysis only and are not intended as leaderboard model inputs.

Attribute Description
Family structure 1=Nuclear, 2=Extended, 3=Single-parent, 4=Blended, 5=Skipped-generation, 6=Other.
Only child status Whether the respondent is an only child (Yes/No).
Parental favoritism If not an only child: 1=Favoring siblings, 2=No favoritism, 3=Favoring self.
Academic performance change Compared with previous semester: 1=Improved, 2=Declined, 3=Stable.
Emotional state change Compared with previous semester: 1=Better, 2=Worse, 3=No change.

DASS-21 Response Scale and Item Encoding

DASS-21 measures Depression, Anxiety, and Stress over the past week. Each item uses a 4-point ordinal response: 0 (did not apply at all), 1 (applied to some degree), 2 (considerable degree), 3 (very much/most of the time).

D/A/S Subscale Reconstruction Rules

Track A2 predicts d01 to d21. Subscale scores are reconstructed with the standard DASS-21 rule:

$$ \begin{aligned} \mathrm{Depression} &= 2 \cdot (d03 + d05 + d10 + d13 + d16 + d17 + d21) \\ \mathrm{Anxiety} &= 2 \cdot (d02 + d04 + d07 + d09 + d15 + d19 + d20) \\ \mathrm{Stress} &= 2 \cdot (d01 + d06 + d08 + d11 + d12 + d14 + d18) \end{aligned} $$

DASS-21 Severity Thresholds (Scaled Scores)

Severity Depression Anxiety Stress
Normal 0-9 0-7 0-14
Mild 10-13 8-9 15-18
Moderate 14-20 10-14 19-25
Severe 21-27 15-19 26-33
Extremely Severe 28+ 20+ 34+

Track A2 Item Paraphrases (D01-D21)

Item Paraphrase Item Paraphrase
D01 Difficulty calming down and settling myself. D02 Mouth felt dry.
D03 Could not experience positive feelings. D04 Breathing discomfort without exertion.
D05 Struggled to get started on tasks. D06 Reacted too strongly to situations.
D07 Experienced trembling. D08 Used a lot of nervous energy.
D09 Worried about panicking or embarrassment. D10 Little to look forward to.
D11 Agitated or unable to stay still. D12 Found it hard to relax.
D13 Felt down-hearted or depressed. D14 Intolerant of interruptions.
D15 Close to breaking down or losing control. D16 Could not become enthusiastic.
D17 Felt not worth much as a person. D18 Easily irritated or touchy.
D19 Unusual heart sensations without exertion. D20 Scared without obvious reason.
D21 Life lacked meaning.

Note: Item text above is paraphrased for reference and aligned with item IDs. For authorized wording, please consult official DASS-21 materials.

Challenge Tracks

Both tracks use the same privacy-preserving feature inputs and official subject-disjoint split protocol. Ranking is generated by the official script for deterministic evaluation.

Track A1: Multi-task Binary Screening (D/A/S)

  • Task: predict three binary targets for Depression, Anxiety, and Stress for each file_id.
  • Output: probabilities \(p_D\), \(p_A\), \(p_S\) in [0,1] for positive class (Mild-or-above).
  • Primary metric: mean F1 across D/A/S with fixed threshold \( \tau = 0.5 \).
  • Secondary metric: mean AUROC across D/A/S.
  • Tie-breaker: higher mean AUROC after primary score tie.
$$ \mathrm{Score}_{A1} = \frac{F1_D + F1_A + F1_S}{3} $$

Submission format: one CSV row per file_id with columns \(p_D\), \(p_A\), \(p_S\).

Track A2: DASS-21 Item Prediction and Reconstruction

  • Task: predict 21 item responses with ordinal labels in {0,1,2,3}.
  • Output: integer columns d01 ... d21 for each file_id.
  • Primary metric: mean Quadratic Weighted Kappa (QWK) over 21 items.
  • Secondary metric: mean MAE over items (lower is better).
  • Auxiliary metric: mean CCC over reconstructed D/A/S scores.
$$ \mathrm{Score}_{A2} = \frac{1}{21}\sum_{i=1}^{21}\mathrm{QWK}(y_i, \hat{y}_i) $$

Ranking uses mean QWK, then mean MAE, then mean CCC as tie-breakers.

Registration

The rankings of our challenge are based on the CodaLab Leaderboard. Entrants will need to register on CodaLab using the GROUP NAME provided on the EULA, or the email used to send the EULA, in order to upload results and view the rankings. We will provide the CodaLab link when the challenge is released.

To further safeguard the security and compliance of the data, please complete the following steps before contacting us to request access to the challenge:

  1. Download the MPDD Dataset License Agreement PDF: User License Agreement (must be signed by a full-time faculty member or researcher; applications not following this rule may be ignored).
  2. Carefully review the agreement. It outlines in detail the usage specifications, restrictions, and the responsibilities and obligations of the licensee. Please read the document thoroughly to ensure complete understanding of all terms and conditions.
  3. Manually sign the agreement by hand after confirming your full understanding and acceptance of the terms, and fill in all required fields.

Once you have completed the above steps, please submit the signed agreement to us via: k3nwong@seu.edu.cn

Baseline & Dataset Release

Coming soon...

Schedule

All deadlines are in AoE (Anywhere on Earth) time.

Data, website, baseline and code available 16 Mar, 2026
Results submission start 09 May, 2026
Results submission deadline 20 May, 2026
Deadline for paper submission 28 May, 2026
Paper acceptance notification 16 Jul, 2026
Deadline for camera-ready papers 06 Aug, 2026

People

Organizers

Haizhou Li

Haizhou Li

Organizer

The Chinese University of Hong Kong, Shenzhen

Shenzhen, China

Hiroshi Ishiguro

Hiroshi Ishiguro

Organizer

Osaka University

Osaka, Japan

Tetsuya Takiguchi

Tetsuya Takiguchi

Organizer

Kobe University

Kobe, Japan

Zhaojie Luo

Zhaojie Luo

Organizer

Southeast University

Nanjing, China

Tomoko Matsui

Tomoko Matsui

Organizer

Shenzhen Loop Area Institute (SLAI)

Shenzhen, China

Kun Qian

Kun Qian

Organizer

Beijing Institute of Technology

Beijing, China

Data Chair

Fei Wang

Fei Wang

Data Chair

Nanjing Medical University

Nanjing, China

Shuqiong Wu

Shuqiong Wu

Data Chair

The University of Osaka

Osaka, Japan

Zhengjun Yue

Zhengjun Yue

Data Chair

Shenzhen Loop Area Institute (SLAI)

Shenzhen, China

Junkun Wang

Junkun Wang

Data Chair

Southeast University

Nanjing, China

Tianhua Qi

Tianhua Qi

Data Chair

Southeast University

Nanjing, China