SemEval-2025 Task 1

AdMIRe - Advancing Multimodal Idiomaticity Representation

This page displays the final official leaderboards of results for the AdMIRe tasks.

Note that participants were required to submit a system description paper to be entered on the final leaderboard. This page will be updated with links to the system description papers when they are published.

Subtask A

Text and Images - English

Team Rank (Test set) Rank (Extended eval set) Top 1 Accuracy (Test set) DCG Score (Test set) Top 1 Accuracy (Extended eval set) DCG Score (Extended eval set)
PALI-NLP 1 1 0.93 3.52 0.83 3.43
dutir914 2 3 0.93 3.46 0.79 3.28
AlexUNLP 3 5 0.93 3.45 0.72 3.22
AIMA 4 10 0.87 3.44 0.48 2.90
daalft 5 2 0.87 3.43 0.81 3.35
PoliTo 6 4 0.87 3.38 0.75 3.20
UCSC NLP T6 7 - 0.87 3.36
Zhoumou 8 6 0.73 3.20 0.69 3.21
HiTZ-Ixa 9 7 0.73 3.13 0.58 3.00
TübingenCL 10 - 0.67 3.16
Howard University-AI4PC 11 - 0.67 3.13
UoR-NCL 12 8 0.67 3.10 0.57 2.96
FJWU_SemEvalSquad 13 11 0.60 2.90 0.47 2.85
JNLP 14 9 0.53 3.14 0.55 3.13
Modgenix 15 - 0.53 2.82
YNU-HPCC 16 - 0.47 2.85
UMUTeam 17 12 0.40 2.68 0.24 2.52

Text Only - English

Team Rank (Test set) Rank (Extended eval set) Top 1 Accuracy (Test set) DCG Score (Test set) Top 1 Accuracy (Extended eval set) DCG Score (Extended eval set)
CTYUN-AI 1 1 0.87 3.51 0.64 3.10
PoliTo 2 2 0.67 3.07 0.59 3.04
daalft 3 5 0.67 3.07 0.33 2.61
JNLP 4 4 0.67 3.04 0.51 2.86
Transformer25 5 3 0.47 2.82 0.54 3.04
ChuenSumi 6 6 0.40 2.89 0.29 2.68

Text and Images - Portuguese

Team Rank (Test set) Rank (Extended eval set) Top 1 Accuracy (Test set) DCG Score (Test set) Top 1 Accuracy (Extended eval set) DCG Score (Extended eval set)
HiTZ-Ixa 1 7 1.00 3.51 0.45 2.82
dutir914 2 2 0.92 3.43 0.69 3.06
Zhoumou 3 3 0.85 3.33 0.67 3.10
daalft 4 4 0.77 3.31 0.56 2.95
PALI-NLP 5 1 0.69 3.21 0.76 3.23
AlexUNLP 6 6 0.62 3.09 0.51 2.91
UoR-NCL 7 5 0.54 3.05 0.56 2.90
YNU-HPCC 8 - 0.38 2.92
UMUTeam 9 8 0.38 2.57 0.18 2.33
HowardUniversity-AI4PC 10 - 0.23 2.64
Modgenix 11 - 0.23 2.57

Text Only - Portuguese

Team Rank (Test set) Rank (Extended eval set) Top 1 Accuracy (Test set) DCG Score (Test set) Top 1 Accuracy (Extended eval set) DCG Score (Extended eval set)
CTYUN-AI 1 1 0.92 3.43 0.56 2.97

Subtask B

Text and Images - English

Team Rank (Test set) Rank (Extended eval set) Image Accuracy (Test set) Sentence Type (Test set) Image Accuracy (Extended eval set) Sentence Type (Extended eval set)
daalft 1 2 0.60 1.00 0.23 0.77
PALI-NLP 2 1 0.60 0.80 0.93 1.00
Modgenix 3 - 0.60 0.60

Text Only - English

Team Rank (Test set) Rank (Extended eval set) Image Accuracy (Test set) Sentence Type (Test set) Image Accuracy (Extended eval set) Sentence Type (Extended eval set)
daalft 1 2 1.00 1.00 0.60 0.77
Transformer25 2 1 0.80 0.60 0.60 0.90