BioASQ Participants Area
Oracle: Results for task B-Phase B
The test results are presented in seperate tables for each type of annotation. The "System Description" of each system is used.
The evaluation measures that are used in Task B are presented
here .
+ Task 1b, Test batch 1
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
Wishart-S1 |
0.9200 |
0.2222 |
0.2778 |
0.2778 |
0.3186 |
0.2147 |
0.2290 |
Wishart-S2 |
0.9200 |
0.2222 |
0.2778 |
0.2778 |
0.3186 |
0.2147 |
0.2290 |
Wishart-S3 |
0.9200 |
0.2222 |
0.3333 |
0.3056 |
0.3067 |
0.2082 |
0.2207 |
main system |
0.3200 |
- | - | - |
0.0037 |
0.0394 |
0.0066 |
BioASQ_Baseline |
0.4400 |
- | - | - |
0.0153 |
0.0402 |
0.0204 |
BioASQ Baseline FS |
0.4800 |
- | - | - |
0.0153 |
0.0402 |
0.0204 |
HPI-S2 |
- |
- | - | - |
- | - | - |
UNCC System 1 |
0.8400 |
- | - | - |
0.0115 |
0.1499 |
0.0209 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
Wishart-S1 |
0.2059 |
0.1463 |
0.2202 |
0.1522 |
Wishart-S2 |
0.2059 |
0.1463 |
0.2202 |
0.1522 |
Wishart-S3 |
0.2059 |
0.1463 |
0.2202 |
0.1522 |
main system |
0.2165 |
0.1231 |
0.2396 |
0.1309 |
BioASQ_Baseline |
0.2266 |
0.0735 |
0.2636 |
0.0843 |
BioASQ Baseline FS |
0.1935 |
0.0628 |
0.2305 |
0.0745 |
HPI-S2 |
0.0872 |
0.1093 |
0.0915 |
0.1155 |
UNCC System 1 |
0.3032 |
0.1266 |
0.3276 |
0.1329 |
+ Task 1b, Test batch 2
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
main system |
0.4231 |
- | - | - |
0.0603 |
0.1040 |
0.0680 |
Wishart-S1 |
0.9615 |
0.2500 |
0.3000 |
0.3000 |
0.4060 |
0.3127 |
0.3336 |
system 2 |
0.4231 |
- | - | - |
0.0437 |
0.0445 |
0.0414 |
system 3 |
0.4231 |
- | - | - |
0.0644 |
0.0440 |
0.0488 |
system 4 |
0.4231 |
- | - | - |
0.0644 |
0.0440 |
0.0488 |
BioASQ_Baseline |
0.2692 |
- | - | - |
0.0612 |
0.2062 |
0.0789 |
BioASQ Baseline FS |
0.5000 |
- | - | - |
0.0612 |
0.2062 |
0.0789 |
UNCC System 1 |
0.8077 |
- | - | - |
0.0065 |
0.1243 |
0.0120 |
lalala |
- |
- | - | - |
- | - | - |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
main system |
0.2122 |
0.0795 |
0.2596 |
0.0890 |
Wishart-S1 |
0.2106 |
0.1197 |
0.2387 |
0.1292 |
system 2 |
0.2204 |
0.0826 |
0.2659 |
0.0899 |
system 3 |
0.2152 |
0.0809 |
0.2592 |
0.0886 |
system 4 |
0.2152 |
0.0809 |
0.2592 |
0.0886 |
BioASQ_Baseline |
0.2074 |
0.0573 |
0.2533 |
0.0675 |
BioASQ Baseline FS |
0.2052 |
0.0563 |
0.2577 |
0.0663 |
UNCC System 1 |
0.3319 |
0.1079 |
0.3596 |
0.1114 |
lalala |
- | - |
+ Task 1b, Test batch 3
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
main system |
0.5769 |
- | - | - |
0.0671 |
0.1013 |
0.0727 |
system 2 |
0.5769 |
- | - | - |
0.0642 |
0.0580 |
0.0575 |
system 3 |
0.5769 |
- | - | - |
0.0594 |
0.0399 |
0.0450 |
BioASQ_Baseline |
0.6538 |
- | - | - |
0.0209 |
0.0635 |
0.0278 |
BioASQ Baseline FS |
0.6154 |
- | - | - |
0.0209 |
0.0635 |
0.0278 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
main system |
0.2563 |
0.0885 |
0.2828 |
0.0912 |
system 2 |
0.2505 |
0.0876 |
0.2806 |
0.0910 |
system 3 |
0.2555 |
0.0884 |
0.2835 |
0.0916 |
BioASQ_Baseline |
0.2670 |
0.0558 |
0.2982 |
0.0596 |
BioASQ Baseline FS |
0.2547 |
0.0534 |
0.2941 |
0.0587 |
+ Task 2b, Test batch 1
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
main system |
0.5938 |
0.0400 |
0.1600 |
0.1000 |
- | - | - |
Biomedical Text Ming |
0.9375 |
0.1600 |
0.1600 |
0.1600 |
0.0572 |
0.0702 |
0.0614 |
SNUMedinfo1 |
- |
- | - | - |
- | - | - |
SNUMedinfo2 |
- |
- | - | - |
- | - | - |
SNUMedinfo3 |
- |
- | - | - |
- | - | - |
SNUMedinfo4 |
- |
- | - | - |
- | - | - |
SNUMedinfo5 |
- |
- | - | - |
- | - | - |
system 2 |
0.9375 |
0.0400 |
0.1600 |
0.1000 |
- | - | - |
Wishart-S3 |
0.8438 |
0.4400 |
0.4800 |
0.4600 |
0.4478 |
0.3335 |
0.3456 |
Wishart-S2 |
0.8438 |
0.4400 |
0.4800 |
0.4600 |
0.4774 |
0.3335 |
0.3621 |
system 3 |
0.9375 |
0.0400 |
0.1600 |
0.1000 |
- | - | - |
BioASQ_Baseline |
0.5313 |
- | - | - |
0.0418 |
0.0766 |
0.0501 |
BioASQ Baseline FS |
0.5000 |
- | - | - |
0.0418 |
0.0766 |
0.0501 |
HPI Master Project |
0.8750 |
- | - | - |
0.0080 |
0.0136 |
0.0084 |
UNCC System 1 |
0.9375 |
- | - | - |
0.0073 |
0.0681 |
0.0130 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
main system |
0.4387 |
0.1381 |
0.4538 |
0.1281 |
Biomedical Text Ming |
0.1724 |
0.1301 |
0.1823 |
0.1278 |
SNUMedinfo1 |
0.4065 |
0.1907 |
0.4100 |
0.1808 |
SNUMedinfo2 |
0.3778 |
0.1855 |
0.3793 |
0.1736 |
SNUMedinfo3 |
0.4971 |
0.1698 |
0.4971 |
0.1562 |
SNUMedinfo4 |
0.3529 |
0.1686 |
0.3542 |
0.1583 |
SNUMedinfo5 |
0.4602 |
0.1588 |
0.4679 |
0.1477 |
system 2 |
0.4400 |
0.1384 |
0.4544 |
0.1282 |
Wishart-S3 |
0.4802 |
0.1724 |
0.4814 |
0.1602 |
Wishart-S2 |
0.4802 |
0.1724 |
0.4814 |
0.1602 |
system 3 |
0.4395 |
0.1383 |
0.4535 |
0.1281 |
BioASQ_Baseline |
0.3741 |
0.0744 |
0.3943 |
0.0727 |
BioASQ Baseline FS |
0.3622 |
0.0758 |
0.3970 |
0.0765 |
HPI Master Project |
0.0222 |
0.0344 |
0.0294 |
0.0444 |
UNCC System 1 |
0.5313 |
0.1454 |
0.5326 |
0.1334 |
+ Task 2b, Test batch 2
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
Biomedical Text Ming |
0.8214 |
- | - | - |
0.1596 |
0.2057 |
0.1618 |
main system |
0.8214 |
0.0435 |
0.1739 |
0.0942 |
- | - | - |
SNUMedinfo2 |
- |
- | - | - |
- | - | - |
SNUMedinfo1 |
- |
- | - | - |
- | - | - |
SNUMedinfo3 |
- |
- | - | - |
- | - | - |
SNUMedinfo4 |
- |
- | - | - |
- | - | - |
SNUMedinfo5 |
- |
- | - | - |
- | - | - |
system 2 |
0.8214 |
0.0435 |
0.1739 |
0.0942 |
- | - | - |
system 3 |
0.8214 |
0.0435 |
0.1739 |
0.0942 |
- | - | - |
system 4 |
0.7500 |
0.0435 |
0.1739 |
0.0942 |
- | - | - |
Wishart-S1 |
0.9286 |
0.1304 |
0.1304 |
0.1304 |
0.4396 |
0.4739 |
0.4122 |
Wishart-S2 |
0.9286 |
0.1304 |
0.1304 |
0.1304 |
0.5120 |
0.4399 |
0.4261 |
system 5 |
0.7143 |
0.0435 |
0.1739 |
0.0942 |
- | - | - |
BioASQ_Baseline |
0.5000 |
- | - | - |
0.1083 |
0.0962 |
0.0951 |
BioASQ Baseline FS |
0.3929 |
- | - | - |
0.1083 |
0.0962 |
0.0951 |
UNCC System 1 |
0.8214 |
- | - | - |
0.0118 |
0.1974 |
0.0219 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
Biomedical Text Ming |
0.1846 |
0.1638 |
0.1948 |
0.1669 |
main system |
0.3280 |
0.1224 |
0.3411 |
0.1225 |
SNUMedinfo2 |
0.3157 |
0.1794 |
0.3198 |
0.1751 |
SNUMedinfo1 |
0.3169 |
0.1829 |
0.3225 |
0.1792 |
SNUMedinfo3 |
0.3116 |
0.1829 |
0.3174 |
0.1794 |
SNUMedinfo4 |
0.2910 |
0.1792 |
0.2973 |
0.1761 |
SNUMedinfo5 |
0.2894 |
0.1677 |
0.2982 |
0.1663 |
system 2 |
0.3247 |
0.1184 |
0.3392 |
0.1186 |
system 3 |
0.3352 |
0.1252 |
0.3493 |
0.1255 |
system 4 |
0.3282 |
0.1219 |
0.3419 |
0.1223 |
Wishart-S1 |
0.3914 |
0.1533 |
0.4089 |
0.1541 |
Wishart-S2 |
0.3914 |
0.1533 |
0.4089 |
0.1541 |
system 5 |
0.3231 |
0.1185 |
0.3385 |
0.1191 |
BioASQ_Baseline |
0.3183 |
0.0813 |
0.3577 |
0.0884 |
BioASQ Baseline FS |
0.3121 |
0.0840 |
0.3359 |
0.0877 |
UNCC System 1 |
0.4075 |
0.1404 |
0.4258 |
0.1399 |
+ Task 2b, Test batch 3
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
Biomedical Text Ming |
0.8333 |
0.0417 |
0.1250 |
0.0833 |
0.1195 |
0.1780 |
0.1373 |
SNUMedinfo1 |
- |
- | - | - |
- | - | - |
SNUMedinfo2 |
- |
- | - | - |
- | - | - |
SNUMedinfo3 |
- |
- | - | - |
- | - | - |
SNUMedinfo4 |
- |
- | - | - |
- | - | - |
SNUMedinfo5 |
- |
- | - | - |
- | - | - |
Wishart-S1 |
0.8889 |
0.0417 |
0.0833 |
0.0556 |
0.4584 |
0.3763 |
0.3909 |
main system |
0.8333 |
- | - | - |
- | - | - |
Asclepius |
0.8333 |
0.0417 |
0.0417 |
0.0417 |
- | - | - |
BioASQ_Baseline |
0.7222 |
- | - | - |
0.1010 |
0.2340 |
0.1114 |
BioASQ Baseline FS |
0.6111 |
- | - | - |
0.1010 |
0.2340 |
0.1114 |
Lab Zhu ,Fdan Univer |
0.8333 |
0.0417 |
0.2083 |
0.0931 |
0.0947 |
0.3322 |
0.1403 |
UNCC System 1 |
0.8333 |
- | - | - |
0.0154 |
0.1522 |
0.0269 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
Biomedical Text Ming |
0.2134 |
0.1564 |
0.2190 |
0.1524 |
SNUMedinfo1 |
0.3382 |
0.1769 |
0.3446 |
0.1723 |
SNUMedinfo2 |
0.3560 |
0.1860 |
0.3605 |
0.1789 |
SNUMedinfo3 |
0.3321 |
0.1793 |
0.3385 |
0.1746 |
SNUMedinfo4 |
0.3307 |
0.1857 |
0.3369 |
0.1808 |
SNUMedinfo5 |
0.3343 |
0.1654 |
0.3407 |
0.1597 |
Wishart-S1 |
0.4331 |
0.1582 |
0.4427 |
0.1525 |
main system |
0.4282 |
0.1552 |
0.4386 |
0.1501 |
Asclepius |
- | - |
BioASQ_Baseline |
0.3319 |
0.0698 |
0.3564 |
0.0711 |
BioASQ Baseline FS |
0.3219 |
0.0630 |
0.3589 |
0.0679 |
Lab Zhu ,Fdan Univer |
0.2046 |
0.2032 |
0.2092 |
0.1963 |
UNCC System 1 |
0.4731 |
0.1662 |
0.4754 |
0.1577 |
+ Task 2b, Test batch 4
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
Hippocrates |
0.8750 |
0.0625 |
0.1875 |
0.1120 |
- | - | - |
Asclepius |
0.8750 |
0.0625 |
0.2188 |
0.1354 |
- | - | - |
Biomedical Text Ming |
0.8750 |
0.0938 |
0.1250 |
0.1042 |
- | - | - |
SNUMedinfo1 |
- |
- | - | - |
- | - | - |
SNUMedinfo2 |
- |
- | - | - |
- | - | - |
SNUMedinfo3 |
- |
- | - | - |
- | - | - |
SNUMedinfo4 |
- |
- | - | - |
- | - | - |
SNUMedinfo5 |
- |
- | - | - |
- | - | - |
Wishart-S1 |
0.9375 |
0.2500 |
0.2813 |
0.2813 |
0.2659 |
0.4029 |
0.2963 |
BioASQ_Baseline |
0.5000 |
- | - | - |
0.1233 |
0.1365 |
0.1062 |
BioASQ Baseline FS |
0.3438 |
- | - | - |
0.1233 |
0.1365 |
0.1062 |
UNCC System 1 |
0.8750 |
- | - | - |
0.0180 |
0.0863 |
0.0286 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
Hippocrates |
- | - |
Asclepius |
- | - |
Biomedical Text Ming |
0.1796 |
0.1518 |
0.1938 |
0.1560 |
SNUMedinfo1 |
0.2963 |
0.1746 |
0.3158 |
0.1788 |
SNUMedinfo2 |
0.2858 |
0.1687 |
0.3087 |
0.1735 |
SNUMedinfo3 |
0.2927 |
0.1743 |
0.3130 |
0.1790 |
SNUMedinfo4 |
0.2756 |
0.1737 |
0.2968 |
0.1794 |
SNUMedinfo5 |
0.3164 |
0.1934 |
0.3384 |
0.1977 |
Wishart-S1 |
0.4072 |
0.1606 |
0.4295 |
0.1596 |
BioASQ_Baseline |
0.3273 |
0.0833 |
0.3677 |
0.0888 |
BioASQ Baseline FS |
0.3075 |
0.0814 |
0.3569 |
0.0874 |
UNCC System 1 |
0.4201 |
0.1404 |
0.4458 |
0.1416 |
+ Task 2b, Test batch 5
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
SNUMedinfo1 |
- |
- | - | - |
- | - | - |
SNUMedinfo2 |
- |
- | - | - |
- | - | - |
SNUMedinfo3 |
- |
- | - | - |
- | - | - |
SNUMedinfo4 |
- |
- | - | - |
- | - | - |
SNUMedinfo5 |
- |
- | - | - |
- | - | - |
Biomedical Text Ming |
1.0000 |
0.1379 |
0.1724 |
0.1466 |
- | - | - |
Asclepius |
0.8333 |
0.0345 |
0.0690 |
0.0460 |
- | - | - |
BioASQ_Baseline |
0.3750 |
- | - | - |
0.1548 |
0.1665 |
0.1319 |
BioASQ Baseline FS |
0.4583 |
- | - | - |
0.1548 |
0.1665 |
0.1319 |
UNCC System 1 |
1.0000 |
- | - | - |
0.0098 |
0.1183 |
0.0177 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
SNUMedinfo1 |
0.2929 |
0.1558 |
0.3178 |
0.1591 |
SNUMedinfo2 |
0.2945 |
0.1575 |
0.3185 |
0.1613 |
SNUMedinfo3 |
0.2895 |
0.1581 |
0.3146 |
0.1616 |
SNUMedinfo4 |
0.2765 |
0.1633 |
0.2984 |
0.1664 |
SNUMedinfo5 |
0.2919 |
0.1509 |
0.3163 |
0.1547 |
Biomedical Text Ming |
0.1685 |
0.1386 |
0.1817 |
0.1429 |
Asclepius |
- | - |
BioASQ_Baseline |
0.3297 |
0.0773 |
0.3599 |
0.0845 |
BioASQ Baseline FS |
0.3366 |
0.0736 |
0.3639 |
0.0827 |
UNCC System 1 |
0.3967 |
0.1429 |
0.4180 |
0.1444 |
+ Task 3b, Test batch 1
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
HPI-S2 |
0.6667 |
- | - | - |
0.0327 |
0.0830 |
0.0424 |
auth-qa-1 |
0.8485 |
0.1154 |
0.1154 |
0.1154 |
- | - | - |
SNUMedinfo1 |
- |
- | - | - |
- | - | - |
SNUMedinfo2 |
- |
- | - | - |
- | - | - |
SNUMedinfo3 |
- |
- | - | - |
- | - | - |
SNUMedinfo4 |
- |
- | - | - |
- | - | - |
SNUMedinfo5 |
- |
- | - | - |
- | - | - |
ilsp.aueb.1 |
- |
- | - | - |
- | - | - |
ilsp.aueb.2 |
- |
- | - | - |
- | - | - |
fdu |
0.8485 |
0.1154 |
0.2692 |
0.1744 |
0.0958 |
0.4312 |
0.1520 |
fdu2 |
0.8485 |
0.1154 |
0.2692 |
0.1744 |
0.0958 |
0.4312 |
0.1520 |
fdu3 |
0.8485 |
0.1154 |
0.2692 |
0.1744 |
0.0931 |
0.4085 |
0.1472 |
fdu4 |
0.8485 |
0.1538 |
0.3077 |
0.2128 |
0.0681 |
0.4998 |
0.1166 |
main system |
0.8485 |
0.1923 |
0.3846 |
0.2641 |
0.1841 |
0.2597 |
0.1987 |
BioASQ_Baseline |
0.4545 |
- | - | - |
- | - | - |
BioASQ Baseline FS |
0.5455 |
- | - | - |
- | - | - |
Lab Zhu ,Fdan Univer |
0.8485 |
0.2308 |
0.3077 |
0.2481 |
0.1255 |
0.5791 |
0.1823 |
Question Summary |
- |
- | - | - |
- | - | - |
test json format |
0.8485 |
- | - | - |
0.0265 |
0.0149 |
0.0170 |
test oracle format |
0.8485 |
- | - | - |
0.1179 |
0.1944 |
0.1408 |
UNCC System 1 |
0.8485 |
0.0769 |
0.1154 |
0.0897 |
0.0122 |
0.2364 |
0.0228 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
HPI-S2 |
0.1897 |
0.1596 |
0.2015 |
0.1636 |
auth-qa-1 |
- | - |
SNUMedinfo1 |
0.3413 |
0.2651 |
0.3550 |
0.2699 |
SNUMedinfo2 |
0.3459 |
0.2699 |
0.3585 |
0.2731 |
SNUMedinfo3 |
0.3386 |
0.2710 |
0.3524 |
0.2747 |
SNUMedinfo4 |
0.2942 |
0.2543 |
0.3076 |
0.2585 |
SNUMedinfo5 |
0.3139 |
0.2658 |
0.3270 |
0.2700 |
ilsp.aueb.1 |
0.3894 |
0.1055 |
0.4144 |
0.1090 |
ilsp.aueb.2 |
0.4183 |
0.1142 |
0.4387 |
0.1165 |
fdu |
0.2801 |
0.2142 |
0.2811 |
0.2098 |
fdu2 |
0.2932 |
0.2142 |
0.3045 |
0.2127 |
fdu3 |
0.2881 |
0.1895 |
0.3076 |
0.1974 |
fdu4 |
0.2881 |
0.1895 |
0.3076 |
0.1974 |
main system |
0.3076 |
0.1473 |
0.3209 |
0.1483 |
BioASQ_Baseline |
0.4053 |
0.1104 |
0.4267 |
0.1133 |
BioASQ Baseline FS |
0.3789 |
0.0994 |
0.3985 |
0.1018 |
Lab Zhu ,Fdan Univer |
0.2490 |
0.1864 |
0.2614 |
0.1864 |
Question Summary |
0.0635 |
0.0950 |
0.0663 |
0.0988 |
test json format |
0.0645 |
0.0396 |
0.0620 |
0.0359 |
test oracle format |
0.0674 |
0.0423 |
0.0653 |
0.0389 |
UNCC System 1 |
0.5240 |
0.2008 |
0.5368 |
0.1964 |
+ Task 3b, Test batch 2
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
main system |
0.8125 |
0.1250 |
0.1875 |
0.1510 |
0.1149 |
0.1798 |
0.1356 |
system 2 |
0.8125 |
0.0938 |
0.1875 |
0.1260 |
0.1149 |
0.1798 |
0.1356 |
system 3 |
0.8125 |
0.0625 |
0.2500 |
0.1391 |
0.1149 |
0.1798 |
0.1356 |
auth-qa-1 |
0.8125 |
0.0313 |
0.0313 |
0.0313 |
0.0357 |
0.0089 |
0.0143 |
ilsp.aueb.1 |
- |
- | - | - |
- | - | - |
ilsp.aueb.2 |
- |
- | - | - |
- | - | - |
HPI-S2 |
0.5625 |
- | - | - |
0.0714 |
0.0161 |
0.0262 |
SNUMedinfo1 |
- |
- | - | - |
- | - | - |
SNUMedinfo2 |
- |
- | - | - |
- | - | - |
SNUMedinfo3 |
- |
- | - | - |
- | - | - |
SNUMedinfo4 |
- |
- | - | - |
- | - | - |
SNUMedinfo5 |
- |
- | - | - |
- | - | - |
fdu4 |
0.8125 |
0.0313 |
0.1563 |
0.0859 |
0.0828 |
0.3772 |
0.1280 |
fdu2 |
0.8125 |
0.0625 |
0.1875 |
0.1172 |
0.0828 |
0.3772 |
0.1280 |
fdu |
0.8125 |
0.0625 |
0.1875 |
0.1172 |
0.0828 |
0.3772 |
0.1280 |
fdu3 |
0.8125 |
0.0625 |
0.1875 |
0.1172 |
0.0829 |
0.3772 |
0.1281 |
fdu5 |
0.7500 |
0.0313 |
0.0313 |
0.0313 |
0.0007 |
0.0083 |
0.0013 |
qaiiit system 1 |
- |
- | - | - |
- | - | - |
BioASQ_Baseline |
0.3125 |
- | - | - |
0.0357 |
0.0089 |
0.0143 |
BioASQ Baseline FS |
0.4375 |
- | - | - |
0.0357 |
0.0089 |
0.0143 |
UNCC System 1 |
- |
- | - | - |
- | - | - |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
main system |
0.2968 |
0.1322 |
0.3055 |
0.1312 |
system 2 |
0.2984 |
0.1326 |
0.3069 |
0.1313 |
system 3 |
0.2983 |
0.1324 |
0.3080 |
0.1314 |
auth-qa-1 |
- | - |
ilsp.aueb.1 |
0.3852 |
0.1043 |
0.4160 |
0.1072 |
ilsp.aueb.2 |
0.3989 |
0.1081 |
0.4368 |
0.1113 |
HPI-S2 |
0.2063 |
0.1664 |
0.2243 |
0.1757 |
SNUMedinfo1 |
0.3953 |
0.2866 |
0.4114 |
0.2847 |
SNUMedinfo2 |
0.3897 |
0.2828 |
0.4073 |
0.2835 |
SNUMedinfo3 |
0.3918 |
0.2909 |
0.4087 |
0.2911 |
SNUMedinfo4 |
0.3509 |
0.2902 |
0.3708 |
0.2923 |
SNUMedinfo5 |
0.3641 |
0.2937 |
0.3848 |
0.2946 |
fdu4 |
0.2604 |
0.1770 |
0.2859 |
0.1850 |
fdu2 |
0.2590 |
0.1761 |
0.2853 |
0.1846 |
fdu |
0.2590 |
0.1761 |
0.2853 |
0.1846 |
fdu3 |
0.2604 |
0.1770 |
0.2859 |
0.1850 |
fdu5 |
- | - |
qaiiit system 1 |
0.3081 |
0.1657 |
0.3353 |
0.1710 |
BioASQ_Baseline |
0.4570 |
0.1251 |
0.4772 |
0.1247 |
BioASQ Baseline FS |
0.4105 |
0.1146 |
0.4388 |
0.1153 |
UNCC System 1 |
0.5451 |
0.1894 |
0.5674 |
0.1852 |
+ Task 3b, Test batch 3
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
HPI-S2 |
0.6207 |
- | - | - |
- | - | - |
SNUMedinfo1 |
- |
- | - | - |
- | - | - |
SNUMedinfo2 |
- |
- | - | - |
- | - | - |
SNUMedinfo3 |
- |
- | - | - |
- | - | - |
SNUMedinfo4 |
- |
- | - | - |
- | - | - |
SNUMedinfo5 |
- |
- | - | - |
- | - | - |
main system |
0.9655 |
0.1923 |
0.2692 |
0.2308 |
0.1529 |
0.1765 |
0.1587 |
system 2 |
0.9655 |
0.0385 |
0.1538 |
0.0897 |
0.1529 |
0.1765 |
0.1587 |
system 3 |
0.9655 |
0.1538 |
0.2692 |
0.1987 |
0.1529 |
0.1765 |
0.1587 |
auth-qa-1 |
0.9655 |
0.0385 |
0.0769 |
0.0577 |
- | - | - |
ilsp.aueb.1 |
- |
- | - | - |
- | - | - |
ilsp.aueb.2 |
- |
- | - | - |
- | - | - |
ilsp.aueb.3 |
- |
- | - | - |
- | - | - |
fdu4 |
0.9655 |
0.0769 |
0.1154 |
0.0962 |
0.0175 |
0.2735 |
0.0324 |
oaqa-3b-3 |
0.0345 |
0.1538 |
0.3462 |
0.2321 |
0.0520 |
0.7383 |
0.0940 |
oaqa-3b-3-e |
0.0345 |
0.1538 |
0.3462 |
0.2321 |
0.0594 |
0.8130 |
0.1072 |
fdu2 |
0.6207 |
0.1154 |
0.1923 |
0.1359 |
0.0851 |
0.3964 |
0.1353 |
fdu3 |
0.6207 |
0.1154 |
0.1923 |
0.1359 |
0.0851 |
0.3964 |
0.1353 |
fdu5 |
0.9655 |
0.0385 |
0.0385 |
0.0385 |
0.0722 |
0.1775 |
0.0390 |
fdu |
0.6207 |
0.1154 |
0.1923 |
0.1359 |
0.1008 |
0.4993 |
0.1608 |
BioASQ_Baseline |
0.3448 |
- | - | - |
- | - | - |
BioASQ Baseline FS |
0.4828 |
- | - | - |
- | - | - |
LabZhu,FDU |
0.9655 |
0.3077 |
0.3077 |
0.3077 |
0.1135 |
0.5679 |
0.1740 |
UNCC System 1 |
0.9655 |
0.0385 |
0.1538 |
0.0865 |
0.0191 |
0.2969 |
0.0341 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
HPI-S2 |
0.2131 |
0.1979 |
0.2360 |
0.2100 |
SNUMedinfo1 |
0.4379 |
0.3338 |
0.4468 |
0.3339 |
SNUMedinfo2 |
0.4363 |
0.3344 |
0.4448 |
0.3343 |
SNUMedinfo3 |
0.4320 |
0.3349 |
0.4407 |
0.3344 |
SNUMedinfo4 |
0.4016 |
0.3349 |
0.4128 |
0.3344 |
SNUMedinfo5 |
0.4121 |
0.3360 |
0.4238 |
0.3359 |
main system |
0.3158 |
0.1459 |
0.3389 |
0.1484 |
system 2 |
0.3154 |
0.1459 |
0.3396 |
0.1490 |
system 3 |
0.3100 |
0.1445 |
0.3341 |
0.1475 |
auth-qa-1 |
- | - |
ilsp.aueb.1 |
0.4573 |
0.1479 |
0.4863 |
0.1486 |
ilsp.aueb.2 |
0.5165 |
0.1651 |
0.5459 |
0.1649 |
ilsp.aueb.3 |
0.4755 |
0.1531 |
0.5072 |
0.1540 |
fdu4 |
- | - |
oaqa-3b-3 |
- | - |
oaqa-3b-3-e |
- | - |
fdu2 |
0.3276 |
0.2714 |
0.3467 |
0.2770 |
fdu3 |
0.3276 |
0.2714 |
0.3467 |
0.2770 |
fdu5 |
- | - |
fdu |
0.3679 |
0.2696 |
0.3933 |
0.2787 |
BioASQ_Baseline |
0.4518 |
0.1455 |
0.4775 |
0.1467 |
BioASQ Baseline FS |
0.4772 |
0.1506 |
0.5005 |
0.1516 |
LabZhu,FDU |
0.2820 |
0.2341 |
0.2868 |
0.2346 |
UNCC System 1 |
0.5994 |
0.2422 |
0.6128 |
0.2364 |
+ Task 3b, Test batch 4
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
SNUMedinfo1 |
- |
- | - | - |
- | - | - |
SNUMedinfo2 |
- |
- | - | - |
- | - | - |
SNUMedinfo3 |
- |
- | - | - |
- | - | - |
SNUMedinfo4 |
- |
- | - | - |
- | - | - |
SNUMedinfo5 |
- |
- | - | - |
- | - | - |
HPI-S2 |
0.5600 |
0.0345 |
0.0345 |
0.0345 |
0.1522 |
0.0473 |
0.0689 |
main system |
0.9600 |
0.2414 |
0.4828 |
0.3374 |
0.1391 |
0.1453 |
0.1335 |
system 2 |
0.9600 |
0.2069 |
0.4483 |
0.3029 |
0.1391 |
0.1453 |
0.1335 |
system 3 |
0.9600 |
0.2414 |
0.3793 |
0.2897 |
0.1391 |
0.1453 |
0.1335 |
ilsp.aueb.1 |
- |
- | - | - |
- | - | - |
ilsp.aueb.2 |
- |
- | - | - |
- | - | - |
ilsp.aueb.3 |
- |
- | - | - |
- | - | - |
oaqa-3b-4 |
0.9600 |
0.4138 |
0.6552 |
0.5098 |
0.3836 |
0.3450 |
0.3155 |
fdu4 |
0.8800 |
0.1724 |
0.3448 |
0.2471 |
0.1446 |
0.5475 |
0.2189 |
fdu |
0.8800 |
0.1724 |
0.3448 |
0.2471 |
0.1446 |
0.5475 |
0.2189 |
fdu2 |
0.9600 |
0.1724 |
0.3448 |
0.2471 |
0.1446 |
0.5475 |
0.2189 |
fdu3 |
0.9600 |
0.1724 |
0.3448 |
0.2471 |
0.1446 |
0.5475 |
0.2189 |
auth-qa-1 |
0.9600 |
0.0690 |
0.2069 |
0.1236 |
- | - | - |
BioASQ_Baseline |
0.3600 |
- | - | - |
- | - | - |
BioASQ Baseline FS |
0.4000 |
- | - | - |
- | - | - |
Lab Zhu ,Fdan Univer |
0.9600 |
0.2414 |
0.4138 |
0.3115 |
0.1232 |
0.4525 |
0.1827 |
LabZhu,FDU |
0.9600 |
0.2414 |
0.4138 |
0.3115 |
0.1348 |
0.4303 |
0.1937 |
LabZhu_FDU |
0.9600 |
0.2414 |
0.4138 |
0.3115 |
0.1467 |
0.4092 |
0.2039 |
LabZhu-FDU |
0.9600 |
0.2414 |
0.3793 |
0.2943 |
0.1630 |
0.4470 |
0.2251 |
UNCC System 1 |
0.9600 |
0.0345 |
0.1379 |
0.0805 |
0.0139 |
0.1570 |
0.0251 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
SNUMedinfo1 |
0.4018 |
0.3304 |
0.4107 |
0.3291 |
SNUMedinfo2 |
0.4250 |
0.3352 |
0.4323 |
0.3327 |
SNUMedinfo3 |
0.3969 |
0.3306 |
0.4036 |
0.3284 |
SNUMedinfo4 |
0.3747 |
0.3282 |
0.3816 |
0.3269 |
SNUMedinfo5 |
0.3879 |
0.3337 |
0.3948 |
0.3320 |
HPI-S2 |
0.2672 |
0.2228 |
0.2872 |
0.2326 |
main system |
0.3430 |
0.1372 |
0.3545 |
0.1375 |
system 2 |
0.3387 |
0.1330 |
0.3509 |
0.1337 |
system 3 |
0.3380 |
0.1318 |
0.3490 |
0.1322 |
ilsp.aueb.1 |
0.4416 |
0.1294 |
0.4631 |
0.1316 |
ilsp.aueb.2 |
0.4672 |
0.1364 |
0.4927 |
0.1387 |
ilsp.aueb.3 |
0.4441 |
0.1306 |
0.4699 |
0.1336 |
oaqa-3b-4 |
- | - |
fdu4 |
0.3260 |
0.2293 |
0.3338 |
0.2295 |
fdu |
0.2490 |
0.2035 |
0.2588 |
0.2041 |
fdu2 |
0.2490 |
0.2035 |
0.2588 |
0.2041 |
fdu3 |
0.3260 |
0.2293 |
0.3338 |
0.2295 |
auth-qa-1 |
- | - |
BioASQ_Baseline |
0.4672 |
0.1375 |
0.4891 |
0.1381 |
BioASQ Baseline FS |
0.4322 |
0.1301 |
0.4495 |
0.1304 |
Lab Zhu ,Fdan Univer |
0.2455 |
0.1993 |
0.2635 |
0.2064 |
LabZhu,FDU |
0.2455 |
0.1993 |
0.2635 |
0.2064 |
LabZhu_FDU |
0.2455 |
0.1993 |
0.2635 |
0.2064 |
LabZhu-FDU |
0.2455 |
0.1993 |
0.2635 |
0.2064 |
UNCC System 1 |
0.5848 |
0.2132 |
0.5950 |
0.2090 |
+ Task 3b, Test batch 5
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
auth-qa-1 |
0.6786 |
0.1364 |
0.1818 |
0.1455 |
- | - | - |
main system |
0.6786 |
0.0455 |
0.1364 |
0.0720 |
0.1139 |
0.1757 |
0.1293 |
system 2 |
0.6786 |
0.0455 |
0.2727 |
0.1303 |
0.1139 |
0.1757 |
0.1293 |
system 3 |
0.6786 |
0.0909 |
0.1364 |
0.1061 |
0.1139 |
0.1757 |
0.1293 |
ilsp.aueb.1 |
- |
- | - | - |
- | - | - |
ilsp.aueb.2 |
- |
- | - | - |
- | - | - |
ilsp.aueb.3 |
- |
- | - | - |
- | - | - |
SNUMedinfo1 |
- |
- | - | - |
- | - | - |
SNUMedinfo2 |
- |
- | - | - |
- | - | - |
SNUMedinfo3 |
- |
- | - | - |
- | - | - |
SNUMedinfo4 |
- |
- | - | - |
- | - | - |
SNUMedinfo5 |
- |
- | - | - |
- | - | - |
fdu |
0.7143 |
0.2273 |
0.2727 |
0.2500 |
0.0914 |
0.4146 |
0.1376 |
fdu2 |
0.6786 |
0.2273 |
0.2727 |
0.2500 |
0.0935 |
0.4146 |
0.1404 |
fdu3 |
0.6786 |
0.2273 |
0.2727 |
0.2500 |
0.0914 |
0.4146 |
0.1376 |
fdu4 |
0.7143 |
0.2273 |
0.2727 |
0.2500 |
0.0914 |
0.4146 |
0.1376 |
HPI-S2 |
0.3571 |
0.0909 |
0.0909 |
0.0909 |
0.0625 |
0.0292 |
0.0397 |
oaqa-3b-4 |
0.6786 |
0.2273 |
0.3182 |
0.2727 |
0.1704 |
0.1139 |
0.1296 |
oaqa-3b-5 |
0.6786 |
0.2273 |
0.3182 |
0.2727 |
0.1643 |
0.2538 |
0.1860 |
YodaQA_base |
0.6786 |
0.1818 |
0.2273 |
0.2045 |
0.1514 |
0.2132 |
0.1580 |
BioASQ_Baseline |
0.7143 |
0.0455 |
0.0455 |
0.0455 |
- | - | - |
BioASQ Baseline FS |
0.6429 |
0.0455 |
0.0455 |
0.0455 |
- | - | - |
Lab Zhu ,Fdan Univer |
0.6786 |
0.2273 |
0.2727 |
0.2500 |
0.0510 |
0.1486 |
0.0720 |
UNCC System 1 |
0.6786 |
0.0909 |
0.1364 |
0.1023 |
0.0016 |
0.0365 |
0.0031 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
auth-qa-1 |
- | - |
main system |
0.3880 |
0.2626 |
0.4000 |
0.2583 |
system 2 |
0.3914 |
0.2639 |
0.4032 |
0.2597 |
system 3 |
0.3901 |
0.2639 |
0.3980 |
0.2593 |
ilsp.aueb.1 |
0.3959 |
0.2084 |
0.4256 |
0.2147 |
ilsp.aueb.2 |
0.3959 |
0.2084 |
0.4256 |
0.2147 |
ilsp.aueb.3 |
0.3705 |
0.1946 |
0.4032 |
0.2024 |
SNUMedinfo1 |
0.3825 |
0.3596 |
0.3899 |
0.3580 |
SNUMedinfo2 |
0.3971 |
0.3643 |
0.4033 |
0.3619 |
SNUMedinfo3 |
0.3780 |
0.3552 |
0.3841 |
0.3532 |
SNUMedinfo4 |
0.3488 |
0.3438 |
0.3559 |
0.3421 |
SNUMedinfo5 |
0.3567 |
0.3505 |
0.3639 |
0.3489 |
fdu |
0.2261 |
0.2234 |
0.2309 |
0.2242 |
fdu2 |
0.2261 |
0.2234 |
0.2309 |
0.2242 |
fdu3 |
0.3172 |
0.2837 |
0.3316 |
0.2860 |
fdu4 |
0.3172 |
0.2837 |
0.3316 |
0.2860 |
HPI-S2 |
0.1629 |
0.1936 |
0.1711 |
0.1945 |
oaqa-3b-4 |
- | - |
oaqa-3b-5 |
- | - |
YodaQA_base |
- | - |
BioASQ_Baseline |
0.4038 |
0.2098 |
0.4313 |
0.2140 |
BioASQ Baseline FS |
0.3540 |
0.1815 |
0.3862 |
0.1895 |
Lab Zhu ,Fdan Univer |
0.3300 |
0.2220 |
0.3336 |
0.2199 |
UNCC System 1 |
0.5651 |
0.3556 |
0.5672 |
0.3438 |
+ Task 4b, Test batch 1
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
auth-qa-1 |
0.9643 |
0.1026 |
0.1538 |
0.1218 |
0.1364 |
0.2636 |
0.1677 |
WS4A |
0.1071 |
0.0256 |
0.0256 |
0.0256 |
0.0844 |
0.1591 |
0.1090 |
HPI-S2 |
0.2500 |
- | - | - |
- | - | - |
UNCC System 1 |
0.9643 |
0.0256 |
0.1026 |
0.0521 |
0.0051 |
0.0909 |
0.0096 |
auth-qa-3 |
0.9643 |
0.0513 |
0.2308 |
0.1128 |
0.1545 |
0.6045 |
0.2394 |
auth-qa-4 |
0.9643 |
0.1026 |
0.2308 |
0.1496 |
0.2364 |
0.4879 |
0.3064 |
auth-qa-2 |
0.9643 |
0.1026 |
0.2308 |
0.1496 |
0.1965 |
0.5818 |
0.2784 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
auth-qa-1 |
- | - |
WS4A |
0.0535 |
0.0674 |
0.0526 |
0.0662 |
HPI-S2 |
0.2354 |
0.1422 |
0.2544 |
0.1498 |
UNCC System 1 |
0.6918 |
0.3180 |
0.7024 |
0.3073 |
auth-qa-3 |
- | - |
auth-qa-4 |
- | - |
auth-qa-2 |
- | - |
+ Task 4b, Test batch 2
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
WS4A |
0.0938 |
- | - | - |
0.0985 |
0.1425 |
0.1023 |
auth-qa-1 |
0.9063 |
0.0645 |
0.1613 |
0.1075 |
0.1429 |
0.1155 |
0.1213 |
Lab Zhu,Fudan Univer |
0.9063 |
0.1935 |
0.2258 |
0.2097 |
0.1106 |
0.3492 |
0.1550 |
LabZhu,FDU |
0.9063 |
0.1935 |
0.2258 |
0.2097 |
0.1031 |
0.3492 |
0.1444 |
HPI-S2 |
- |
- | - | - |
- | - | - |
LabZhu_FDU |
0.0938 |
0.1935 |
0.2258 |
0.2097 |
0.1031 |
0.3492 |
0.1444 |
Lab Zhu ,Fdan Univer |
0.9063 |
0.1935 |
0.2258 |
0.2097 |
0.1052 |
0.3492 |
0.1474 |
LabZhu-FDU |
0.9063 |
0.1935 |
0.2581 |
0.2258 |
0.1038 |
0.3492 |
0.1454 |
UNCC System 1 |
0.9063 |
0.0323 |
0.0645 |
0.0403 |
0.0311 |
0.2525 |
0.0509 |
auth-qa-2 |
0.9063 |
0.0968 |
0.3226 |
0.1909 |
0.1524 |
0.3055 |
0.1926 |
auth-qa-4 |
0.9063 |
0.0968 |
0.3226 |
0.1935 |
0.1524 |
0.1506 |
0.1433 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
WS4A |
0.0322 |
0.0442 |
0.0300 |
0.0420 |
auth-qa-1 |
- | - |
Lab Zhu,Fudan Univer |
0.3354 |
0.3068 |
0.3384 |
0.3063 |
LabZhu,FDU |
0.3354 |
0.3068 |
0.3384 |
0.3063 |
HPI-S2 |
0.2583 |
0.1380 |
0.2910 |
0.1477 |
LabZhu_FDU |
0.3354 |
0.3068 |
0.3384 |
0.3063 |
Lab Zhu ,Fdan Univer |
0.3354 |
0.3068 |
0.3384 |
0.3063 |
LabZhu-FDU |
0.3354 |
0.3068 |
0.3384 |
0.3063 |
UNCC System 1 |
0.6777 |
0.2829 |
0.6897 |
0.2742 |
auth-qa-2 |
- | - |
auth-qa-4 |
- | - |
+ Task 4b, Test batch 3
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
auth-qa-1 |
0.9600 |
0.1154 |
0.1923 |
0.1442 |
0.2500 |
0.2873 |
0.2580 |
Lab Zhu ,Fdan Univer |
0.9600 |
0.1923 |
0.2692 |
0.2192 |
0.1450 |
0.5929 |
0.2181 |
LabZhu,FDU |
0.9600 |
0.1923 |
0.2692 |
0.2192 |
0.1444 |
0.6214 |
0.2176 |
oaqa-3b-3 |
0.5200 |
0.2308 |
0.2692 |
0.2436 |
0.5396 |
0.5008 |
0.4828 |
LabZhu_FDU |
0.9600 |
0.1923 |
0.2692 |
0.2192 |
0.1420 |
0.5929 |
0.2132 |
LabZhu-FDU |
0.0400 |
0.1923 |
0.2692 |
0.2192 |
0.1420 |
0.5929 |
0.2132 |
Lab Zhu,Fudan Univer |
0.9600 |
0.1923 |
0.2692 |
0.2192 |
0.1455 |
0.5770 |
0.2185 |
HPI-S2 |
- |
- | - | - |
- | - | - |
WS4A |
0.2400 |
0.0385 |
0.0385 |
0.0385 |
0.1172 |
0.2817 |
0.1609 |
UNCC System 1 |
0.9600 |
- | - | - |
0.0417 |
0.3667 |
0.0729 |
auth-qa-2 |
0.9600 |
0.1923 |
0.3462 |
0.2596 |
0.2048 |
0.6135 |
0.3001 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
auth-qa-1 |
- | - |
Lab Zhu ,Fdan Univer |
0.3312 |
0.3041 |
0.3420 |
0.3049 |
LabZhu,FDU |
0.3312 |
0.3041 |
0.3420 |
0.3049 |
oaqa-3b-3 |
- | - |
LabZhu_FDU |
0.3312 |
0.3041 |
0.3420 |
0.3049 |
LabZhu-FDU |
0.3312 |
0.3041 |
0.3420 |
0.3049 |
Lab Zhu,Fudan Univer |
0.3312 |
0.3041 |
0.3420 |
0.3049 |
HPI-S2 |
0.2903 |
0.1712 |
0.3146 |
0.1782 |
WS4A |
0.0874 |
0.1135 |
0.0857 |
0.1120 |
UNCC System 1 |
0.6364 |
0.2978 |
0.6527 |
0.2879 |
auth-qa-2 |
- | - |
+ Task 4b, Test batch 4
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
Lab Zhu ,Fdan Univer |
0.9524 |
0.0968 |
0.1935 |
0.1371 |
0.1970 |
0.2258 |
0.1711 |
Lab Zhu,Fudan Univer |
0.9524 |
0.0968 |
0.1935 |
0.1371 |
0.2169 |
0.2925 |
0.1986 |
HPI-S2 |
- |
- | - | - |
- | - | - |
oaqa-3b-4 |
0.6667 |
0.2903 |
0.3871 |
0.3253 |
0.2478 |
0.5494 |
0.3115 |
LabZhu,FDU |
0.9524 |
0.0968 |
0.1935 |
0.1371 |
0.2095 |
0.2925 |
0.1908 |
auth-qa-1 |
0.9524 |
0.0323 |
0.1935 |
0.0806 |
0.0333 |
0.0444 |
0.0381 |
Oaqa5b-tfidf |
- |
- | - | - |
- | - | - |
UNCC System 1 |
0.9524 |
0.0323 |
0.0323 |
0.0323 |
0.0162 |
0.2717 |
0.0293 |
auth-qa-2 |
0.9524 |
0.2903 |
0.4194 |
0.3414 |
0.2067 |
0.5887 |
0.2868 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
Lab Zhu ,Fdan Univer |
0.3363 |
0.3145 |
0.3308 |
0.3064 |
Lab Zhu,Fudan Univer |
0.3363 |
0.3145 |
0.3308 |
0.3064 |
HPI-S2 |
0.2525 |
0.1618 |
0.2701 |
0.1666 |
oaqa-3b-4 |
- | - |
LabZhu,FDU |
0.3363 |
0.3145 |
0.3308 |
0.3064 |
auth-qa-1 |
- | - |
Oaqa5b-tfidf |
0.6658 |
0.2626 |
0.6606 |
0.2524 |
UNCC System 1 |
0.7196 |
0.3095 |
0.7177 |
0.2960 |
auth-qa-2 |
- | - |
+ Task 4b, Test batch 5
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
auth-qa-1 |
1.0000 |
0.0606 |
0.1212 |
0.0833 |
0.0875 |
0.0830 |
0.0835 |
Lab Zhu ,Fdan Univer |
1.0000 |
0.1818 |
0.3333 |
0.2475 |
0.0938 |
0.3479 |
0.1417 |
Lab Zhu,Fudan Univer |
1.0000 |
0.2121 |
0.3030 |
0.2475 |
0.1596 |
0.2188 |
0.1493 |
HPI-S2 |
- |
- | - | - |
- | - | - |
oaqa-3b-5 |
0.7407 |
0.2121 |
0.3939 |
0.2854 |
0.2662 |
0.4170 |
0.2897 |
oaqa-3b-5-e |
1.0000 |
0.2121 |
0.3939 |
0.2854 |
0.2165 |
0.3754 |
0.2597 |
LabZhu,FDU |
1.0000 |
0.2121 |
0.3030 |
0.2475 |
0.1636 |
0.2111 |
0.1507 |
LabZhu_FDU |
1.0000 |
0.1818 |
0.3333 |
0.2475 |
0.1036 |
0.3302 |
0.1502 |
LabZhu-FDU |
1.0000 |
0.1818 |
0.3333 |
0.2475 |
0.0854 |
0.3579 |
0.1328 |
WS4A |
0.2593 |
- | - | - |
0.0589 |
0.0698 |
0.0560 |
test oracle format |
1.0000 |
- | - | - |
0.1715 |
0.2175 |
0.1795 |
UNCC System 1 |
1.0000 |
0.0606 |
0.0909 |
0.0707 |
0.0321 |
0.2164 |
0.0545 |
auth-qa-2 |
1.0000 |
0.1515 |
0.3636 |
0.2273 |
0.1750 |
0.4065 |
0.2304 |
summary |
- |
- | - | - |
- | - | - |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
auth-qa-1 |
- | - |
Lab Zhu ,Fdan Univer |
0.4762 |
0.4734 |
0.4711 |
0.4645 |
Lab Zhu,Fudan Univer |
0.4762 |
0.4734 |
0.4711 |
0.4645 |
HPI-S2 |
0.3541 |
0.2067 |
0.3799 |
0.2114 |
oaqa-3b-5 |
- | - |
oaqa-3b-5-e |
- | - |
LabZhu,FDU |
0.4762 |
0.4734 |
0.4711 |
0.4645 |
LabZhu_FDU |
0.4762 |
0.4734 |
0.4711 |
0.4645 |
LabZhu-FDU |
0.4762 |
0.4734 |
0.4711 |
0.4645 |
WS4A |
0.0625 |
0.0786 |
0.0583 |
0.0733 |
test oracle format |
0.0192 |
0.0198 |
0.0183 |
0.0189 |
UNCC System 1 |
0.7266 |
0.3259 |
0.7250 |
0.3130 |
auth-qa-2 |
- | - |
summary |
0.5011 |
0.6105 |
0.4882 |
0.5984 |
+ Task 5b, Test batch 1
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
MQ-1 |
0.8824 |
- | - | - |
- | - | - |
MQ-2 |
0.8824 |
- | - | - |
- | - | - |
Deep QA (single) |
- |
0.5200 |
0.6000 |
0.5600 |
0.3269 |
0.4393 |
0.3365 |
Deep QA (ensemble) |
- |
0.5600 |
0.6800 |
0.6033 |
0.3106 |
0.4164 |
0.3341 |
auth-qa-1 |
0.8824 |
0.1200 |
0.1200 |
0.1200 |
- | - | - |
sarrouti |
0.7647 |
0.1200 |
0.2800 |
0.1833 |
0.1909 |
0.2539 |
0.2073 |
Olelo-GS |
0.7647 |
0.0400 |
0.0400 |
0.0400 |
0.0402 |
0.1061 |
0.0477 |
Olelo |
0.8235 |
0.0400 |
0.0400 |
0.0400 |
0.0193 |
0.0492 |
0.0240 |
HPI_SRL |
0.8824 |
- | - | - |
0.0045 |
0.0027 |
0.0034 |
limsi-reader-UMLS-r1 |
0.6471 |
0.0800 |
0.2000 |
0.1267 |
0.0523 |
0.0508 |
0.0500 |
MQ-3 |
0.8824 |
- | - | - |
- | - | - |
MQ-4 |
0.8824 |
- | - | - |
- | - | - |
Lab Zhu ,Fdan Univer |
0.8824 |
0.4000 |
0.4400 |
0.4200 |
0.1489 |
0.4619 |
0.2068 |
SemanticRoleLabeling |
0.8824 |
- | - | - |
0.1119 |
0.1301 |
0.1158 |
Lab Zhu,Fudan Univer |
0.8824 |
0.4000 |
0.4400 |
0.4200 |
0.1467 |
0.3782 |
0.1965 |
BioASQ_Baseline |
0.3529 |
0.2800 |
0.4000 |
0.3333 |
0.2658 |
0.4715 |
0.3103 |
auth-qa-2 |
0.8824 |
0.3600 |
0.5200 |
0.4200 |
0.1682 |
0.4869 |
0.2399 |
auth-qa-3 |
0.8824 |
0.3600 |
0.4800 |
0.4133 |
0.2337 |
0.4749 |
0.3004 |
UNCC System 2 |
0.8824 |
0.1200 |
0.1600 |
0.1333 |
0.0241 |
0.3252 |
0.0441 |
124M FT |
- |
- | - | - |
- | - | - |
system of teamdaiict |
- |
- | - | - |
- | - | - |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
MQ-1 |
0.5470 |
0.3428 |
0.5599 |
0.3367 |
MQ-2 |
0.5131 |
0.3211 |
0.5221 |
0.3135 |
Deep QA (single) |
- | - |
Deep QA (ensemble) |
- | - |
auth-qa-1 |
- | - |
sarrouti |
0.5087 |
0.3600 |
0.5247 |
0.3577 |
Olelo-GS |
0.3081 |
0.1444 |
0.3362 |
0.1510 |
Olelo |
0.2536 |
0.0856 |
0.3004 |
0.0929 |
HPI_SRL |
0.0608 |
0.0613 |
0.0653 |
0.0640 |
limsi-reader-UMLS-r1 |
- | - |
MQ-3 |
0.5194 |
0.2605 |
0.5334 |
0.2571 |
MQ-4 |
0.4136 |
0.2031 |
0.4333 |
0.2035 |
Lab Zhu ,Fdan Univer |
0.3328 |
0.3181 |
0.3401 |
0.3211 |
SemanticRoleLabeling |
0.0933 |
0.0925 |
0.0974 |
0.0946 |
Lab Zhu,Fudan Univer |
0.3328 |
0.3181 |
0.3401 |
0.3211 |
BioASQ_Baseline |
- | - |
auth-qa-2 |
- | - |
auth-qa-3 |
- | - |
UNCC System 2 |
0.6716 |
0.3130 |
0.6712 |
0.2998 |
124M FT |
- | - |
system of teamdaiict |
- | - |
+ Task 5b, Test batch 2
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
auth-qa-1 |
0.9630 |
0.1290 |
0.1935 |
0.1559 |
0.1133 |
0.3989 |
0.1731 |
Olelo |
0.7037 |
0.0323 |
0.0645 |
0.0430 |
0.0217 |
0.0522 |
0.0287 |
Olelo-GS |
0.9630 |
0.0323 |
0.0323 |
0.0323 |
0.0222 |
0.0389 |
0.0281 |
Lab Zhu ,Fdan Univer |
0.9630 |
0.3226 |
0.4194 |
0.3710 |
0.3828 |
0.6478 |
0.4597 |
sarrouti |
0.7778 |
0.0323 |
0.1935 |
0.0887 |
0.2400 |
0.3922 |
0.2920 |
Lab Zhu,Fudan Univer |
0.9630 |
0.4516 |
0.5161 |
0.4839 |
0.4178 |
0.6700 |
0.5001 |
LabZhu,FDU |
0.9630 |
0.2903 |
0.4839 |
0.3570 |
0.2555 |
0.5144 |
0.3249 |
Deep QA (single) |
- |
0.3226 |
0.5161 |
0.4086 |
0.2571 |
0.3756 |
0.2895 |
Deep QA (ensemble) |
- |
0.3871 |
0.5161 |
0.4419 |
0.2530 |
0.2978 |
0.2617 |
MQ-1 |
0.9630 |
- | - | - |
- | - | - |
MQ-2 |
0.9630 |
- | - | - |
- | - | - |
MQ-3 |
0.9630 |
- | - | - |
- | - | - |
MQ-4 |
0.9630 |
- | - | - |
- | - | - |
Oaqa5b-tfidf |
- |
- | - | - |
- | - | - |
Oaqa5b |
- |
- | - | - |
- | - | - |
SemanticRoleLabeling |
0.9630 |
- | - | - |
0.0952 |
0.1422 |
0.1123 |
LabZhu_FDU |
0.9630 |
0.3226 |
0.4839 |
0.3839 |
0.3561 |
0.6344 |
0.4396 |
limsi-reader-UMLS-r1 |
0.5556 |
0.0968 |
0.1290 |
0.1075 |
0.0307 |
0.1133 |
0.0462 |
Oaqa 5b |
- |
- | - | - |
- | - | - |
LabZhu-FDU |
0.9630 |
0.2258 |
0.3226 |
0.2688 |
0.2572 |
0.4789 |
0.3211 |
limsi-reader-UMLS-r2 |
0.5556 |
0.0323 |
0.1290 |
0.0656 |
0.0329 |
0.0933 |
0.0455 |
BioASQ_Baseline |
0.3704 |
0.1613 |
0.3548 |
0.2215 |
0.2704 |
0.4433 |
0.2931 |
auth-qa-3 |
0.9630 |
0.1935 |
0.3871 |
0.2796 |
0.3161 |
0.6078 |
0.3949 |
auth-qa-2 |
0.8519 |
- | - | - |
- | - | - |
UNCC System 2 |
0.9630 |
0.0323 |
0.1290 |
0.0575 |
0.0353 |
0.2700 |
0.0600 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
auth-qa-1 |
0.2600 |
0.2483 |
0.2736 |
0.2534 |
Olelo |
0.2807 |
0.1253 |
0.3036 |
0.1299 |
Olelo-GS |
0.2151 |
0.0495 |
0.2590 |
0.0578 |
Lab Zhu ,Fdan Univer |
0.3364 |
0.3197 |
0.3372 |
0.3154 |
sarrouti |
0.4823 |
0.3250 |
0.4828 |
0.3180 |
Lab Zhu,Fudan Univer |
0.3379 |
0.3203 |
0.3393 |
0.3161 |
LabZhu,FDU |
0.3354 |
0.3186 |
0.3349 |
0.3136 |
Deep QA (single) |
- | - |
Deep QA (ensemble) |
- | - |
MQ-1 |
0.5117 |
0.3384 |
0.5167 |
0.3307 |
MQ-2 |
0.5351 |
0.3453 |
0.5384 |
0.3358 |
MQ-3 |
0.4859 |
0.2763 |
0.4993 |
0.2705 |
MQ-4 |
0.3895 |
0.2010 |
0.4074 |
0.1999 |
Oaqa5b-tfidf |
0.1332 |
0.1026 |
0.1352 |
0.1016 |
Oaqa5b |
0.1939 |
0.1281 |
0.1928 |
0.1232 |
SemanticRoleLabeling |
0.0506 |
0.0511 |
0.0499 |
0.0506 |
LabZhu_FDU |
0.3438 |
0.3271 |
0.3453 |
0.3229 |
limsi-reader-UMLS-r1 |
- | - |
Oaqa 5b |
0.1923 |
0.1259 |
0.1924 |
0.1221 |
LabZhu-FDU |
0.3362 |
0.3196 |
0.3356 |
0.3147 |
limsi-reader-UMLS-r2 |
- | - |
BioASQ_Baseline |
- | - |
auth-qa-3 |
- | - |
auth-qa-2 |
- | - |
UNCC System 2 |
0.6918 |
0.3174 |
0.6903 |
0.3030 |
+ Task 5b, Test batch 3
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
Olelo |
0.6774 |
- | - | - |
0.0410 |
0.1044 |
0.0548 |
Olelo-GS |
0.7742 |
- | - | - |
0.0281 |
0.1178 |
0.0423 |
auth-qa-1 |
0.8065 |
0.1538 |
0.3077 |
0.2147 |
0.1000 |
0.3435 |
0.1479 |
MQ-1 |
0.8065 |
- | - | - |
- | - | - |
MQ-2 |
0.8065 |
- | - | - |
- | - | - |
limsi-reader-UMLS-r2 |
0.5161 |
0.0385 |
0.0769 |
0.0481 |
0.0310 |
0.2222 |
0.0530 |
limsi-reader-UMLS-r1 |
0.5161 |
- | - | - |
0.0067 |
0.0667 |
0.0121 |
Lab Zhu,Fudan Univer |
0.8065 |
0.3462 |
0.4231 |
0.3846 |
0.3564 |
0.4800 |
0.3874 |
Lab Zhu ,Fdan Univer |
0.8065 |
0.3077 |
0.3462 |
0.3269 |
0.3535 |
0.4800 |
0.3852 |
MQ-3 |
0.8065 |
- | - | - |
- | - | - |
MQ-4 |
0.8065 |
- | - | - |
- | - | - |
sarrouti |
0.8387 |
0.1923 |
0.2692 |
0.2212 |
0.2000 |
0.4117 |
0.2625 |
SemanticRoleLabeling |
0.8065 |
- | - | - |
0.1456 |
0.2524 |
0.1715 |
LabZhu,FDU |
0.8065 |
0.2308 |
0.3077 |
0.2577 |
0.0944 |
0.2244 |
0.1247 |
Deep QA (single) |
0.8065 |
0.3077 |
0.5769 |
0.4308 |
0.3643 |
0.5670 |
0.4128 |
Deep QA (ensemble) |
0.8065 |
0.3077 |
0.6154 |
0.4212 |
0.4467 |
0.5981 |
0.4925 |
Oaqa5b-tfidf |
0.5806 |
0.1154 |
0.2692 |
0.1923 |
0.1897 |
0.5743 |
0.2623 |
LabZhu_FDU |
0.8065 |
0.3462 |
0.4231 |
0.3846 |
0.3580 |
0.4800 |
0.3884 |
LabZhu-FDU |
0.8065 |
0.2308 |
0.3077 |
0.2577 |
0.0889 |
0.2244 |
0.1169 |
Oaqa-5b |
0.5806 |
0.2692 |
0.3846 |
0.3173 |
0.1547 |
0.6371 |
0.2375 |
BioASQ_Baseline |
0.5161 |
0.1154 |
0.2692 |
0.1923 |
0.1723 |
0.5632 |
0.2506 |
auth-qa-2 |
0.8065 |
0.3462 |
0.4615 |
0.4038 |
0.1800 |
0.6171 |
0.2661 |
UNCC System 2 |
0.8065 |
0.0385 |
0.0769 |
0.0462 |
0.0195 |
0.3673 |
0.0367 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
Olelo |
0.3364 |
0.1602 |
0.3534 |
0.1642 |
Olelo-GS |
0.2879 |
0.0769 |
0.3239 |
0.0855 |
auth-qa-1 |
0.2788 |
0.2668 |
0.2973 |
0.2762 |
MQ-1 |
0.5771 |
0.3875 |
0.5813 |
0.3797 |
MQ-2 |
0.6062 |
0.4051 |
0.6085 |
0.3955 |
limsi-reader-UMLS-r2 |
- | - |
limsi-reader-UMLS-r1 |
- | - |
Lab Zhu,Fudan Univer |
0.3279 |
0.3272 |
0.3303 |
0.3246 |
Lab Zhu ,Fdan Univer |
0.3279 |
0.3272 |
0.3299 |
0.3243 |
MQ-3 |
0.5583 |
0.3189 |
0.5767 |
0.3167 |
MQ-4 |
0.5002 |
0.2677 |
0.5158 |
0.2652 |
sarrouti |
0.5658 |
0.3829 |
0.5729 |
0.3775 |
SemanticRoleLabeling |
0.0482 |
0.0521 |
0.0484 |
0.0526 |
LabZhu,FDU |
0.3259 |
0.3323 |
0.3261 |
0.3287 |
Deep QA (single) |
- | - |
Deep QA (ensemble) |
- | - |
Oaqa5b-tfidf |
0.2010 |
0.1272 |
0.1993 |
0.1244 |
LabZhu_FDU |
0.3279 |
0.3272 |
0.3303 |
0.3246 |
LabZhu-FDU |
0.3259 |
0.3323 |
0.3261 |
0.3287 |
Oaqa-5b |
0.2005 |
0.1268 |
0.1982 |
0.1234 |
BioASQ_Baseline |
- | - |
auth-qa-2 |
- | - |
UNCC System 2 |
0.7802 |
0.3649 |
0.7769 |
0.3519 |
+ Task 5b, Test batch 4
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
auth-qa-1 |
0.5517 |
0.1212 |
0.2424 |
0.1551 |
0.0564 |
0.1333 |
0.0718 |
Deep QA (single) |
0.5517 |
0.2424 |
0.4242 |
0.3025 |
0.2254 |
0.3564 |
0.2419 |
Deep QA (ensemble) |
0.5517 |
0.3333 |
0.5455 |
0.4162 |
0.2833 |
0.3436 |
0.2927 |
MQ-1 |
0.5517 |
- | - | - |
- | - | - |
MQ-2 |
0.5517 |
- | - | - |
- | - | - |
auth-qa-2 |
0.5517 |
0.0303 |
0.0606 |
0.0404 |
- | - | - |
sarrouti |
0.6207 |
0.0909 |
0.1212 |
0.0970 |
0.1077 |
0.2013 |
0.1369 |
Lab Zhu ,Fdan Univer |
0.5517 |
0.1212 |
0.2424 |
0.1742 |
0.1143 |
0.3077 |
0.1599 |
auth-qa-3 |
0.5517 |
0.0303 |
0.0909 |
0.0465 |
- | - | - |
Lab Zhu,Fudan Univer |
0.5517 |
0.1818 |
0.3636 |
0.2601 |
0.3608 |
0.4231 |
0.3752 |
Olelo |
0.5517 |
- | - | - |
0.0513 |
0.0513 |
0.0513 |
Olelo-GS |
0.5172 |
- | - | - |
0.0513 |
0.0513 |
0.0513 |
limsi-reader-UMLS-r1 |
0.5172 |
- | - | - |
0.0192 |
0.0513 |
0.0280 |
Oaqa 5b |
0.6207 |
0.0909 |
0.1212 |
0.1061 |
0.1165 |
0.4615 |
0.1792 |
Oaqa-5b |
0.6552 |
0.1515 |
0.2424 |
0.1970 |
0.1252 |
0.5353 |
0.1909 |
SemanticRoleLabeling |
0.5517 |
0.0303 |
0.0606 |
0.0379 |
0.0846 |
0.1122 |
0.0943 |
limsi-reader-UMLS-r2 |
0.5172 |
0.0303 |
0.0303 |
0.0303 |
0.0371 |
0.1667 |
0.0504 |
L2PS - DeepQA |
0.5172 |
- | - | - |
0.0207 |
0.2423 |
0.0338 |
Oaqa5b-tfidf |
0.6207 |
0.0909 |
0.1212 |
0.1061 |
0.1165 |
0.4615 |
0.1792 |
Oaqa5b |
- |
- | - | - |
- | - | - |
LabZhu,FDU |
0.5517 |
0.2424 |
0.4242 |
0.3207 |
0.3608 |
0.4231 |
0.3752 |
MQ-4 |
0.5517 |
- | - | - |
- | - | - |
LabZhu_FDU |
0.5517 |
0.2727 |
0.4545 |
0.3510 |
0.3608 |
0.4231 |
0.3752 |
MQ-3 |
0.5517 |
- | - | - |
- | - | - |
Basic QA pipline |
0.5517 |
0.0909 |
0.2424 |
0.1414 |
0.0769 |
0.1462 |
0.0967 |
LabZhu-FDU |
0.5517 |
0.1212 |
0.2121 |
0.1667 |
0.1239 |
0.3077 |
0.1692 |
BioASQ_Baseline |
0.4828 |
0.0303 |
0.1212 |
0.0682 |
0.1624 |
0.4276 |
0.2180 |
auth-qa-2 |
0.6207 |
- | - | - |
- | - | - |
UNCC System 2 |
0.5517 |
0.0303 |
0.0909 |
0.0455 |
0.0250 |
0.2051 |
0.0389 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
auth-qa-1 |
0.3473 |
0.3548 |
0.3543 |
0.3561 |
Deep QA (single) |
- | - |
Deep QA (ensemble) |
- | - |
MQ-1 |
0.5617 |
0.4069 |
0.5533 |
0.3954 |
MQ-2 |
0.5822 |
0.4260 |
0.5742 |
0.4126 |
auth-qa-2 |
- | - |
sarrouti |
0.5667 |
0.4238 |
0.5640 |
0.4122 |
Lab Zhu ,Fdan Univer |
0.3564 |
0.3470 |
0.3535 |
0.3397 |
auth-qa-3 |
- | - |
Lab Zhu,Fudan Univer |
0.3693 |
0.3516 |
0.3665 |
0.3443 |
Olelo |
0.2434 |
0.0774 |
0.2851 |
0.0840 |
Olelo-GS |
0.3398 |
0.1899 |
0.3462 |
0.1880 |
limsi-reader-UMLS-r1 |
- | - |
Oaqa 5b |
0.5685 |
0.4044 |
0.5619 |
0.3932 |
Oaqa-5b |
0.6704 |
0.4132 |
0.6631 |
0.4011 |
SemanticRoleLabeling |
0.0804 |
0.0826 |
0.0744 |
0.0770 |
limsi-reader-UMLS-r2 |
- | - |
L2PS - DeepQA |
- | - |
Oaqa5b-tfidf |
0.6766 |
0.4163 |
0.6692 |
0.4024 |
Oaqa5b |
0.6726 |
0.4157 |
0.6642 |
0.4027 |
LabZhu,FDU |
0.3793 |
0.3600 |
0.3765 |
0.3526 |
MQ-4 |
0.4251 |
0.2237 |
0.4367 |
0.2219 |
LabZhu_FDU |
0.3851 |
0.3642 |
0.3822 |
0.3568 |
MQ-3 |
0.5352 |
0.3131 |
0.5313 |
0.3029 |
Basic QA pipline |
0.0803 |
0.0860 |
0.0794 |
0.0839 |
LabZhu-FDU |
0.3564 |
0.3470 |
0.3535 |
0.3397 |
BioASQ_Baseline |
- | - |
auth-qa-2 |
- | - |
UNCC System 2 |
0.7363 |
0.4193 |
0.7258 |
0.4051 |
+ Task 5b, Test batch 5
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
auth-qa-1 |
0.4615 |
0.1429 |
0.2286 |
0.1762 |
0.2000 |
0.4289 |
0.2418 |
Olelo |
0.5000 |
- | - | - |
0.0331 |
0.0826 |
0.0379 |
Olelo-GS |
0.4615 |
- | - | - |
0.0147 |
0.0583 |
0.0202 |
auth-qa-2 |
0.4615 |
- | - | - |
- | - | - |
Deep QA (single) |
0.4615 |
0.3714 |
0.4571 |
0.3924 |
0.3308 |
0.4990 |
0.3606 |
Deep QA (ensemble) |
0.4615 |
0.2571 |
0.5143 |
0.3510 |
0.3926 |
0.4854 |
0.3911 |
BioASQ_Baseline |
0.6923 |
0.0571 |
0.2000 |
0.1167 |
0.2898 |
0.5585 |
0.3415 |
Lab Zhu ,Fdan Univer |
0.4615 |
0.2571 |
0.3714 |
0.3010 |
0.2338 |
0.5289 |
0.2574 |
limsi-reader-UMLS-r2 |
0.5385 |
0.0286 |
0.0857 |
0.0452 |
0.0275 |
0.2178 |
0.0474 |
sarrouti |
0.4615 |
0.1714 |
0.2857 |
0.2071 |
0.2182 |
0.3178 |
0.2529 |
MQ-1 |
0.4615 |
- | - | - |
- | - | - |
MQ-2 |
0.4615 |
- | - | - |
- | - | - |
MQ-3 |
0.4615 |
- | - | - |
- | - | - |
MQ-4 |
0.4615 |
- | - | - |
- | - | - |
Oaqa-5b |
0.6154 |
0.1429 |
0.2286 |
0.1810 |
0.1280 |
0.6109 |
0.1970 |
Oaqa5b |
0.6154 |
0.2000 |
0.3143 |
0.2381 |
0.2095 |
0.6433 |
0.2660 |
Oaqa 5b |
0.6154 |
0.2000 |
0.3143 |
0.2381 |
0.2095 |
0.6433 |
0.2660 |
oaqa5b5 |
0.6154 |
0.2000 |
0.3143 |
0.2381 |
0.2095 |
0.6433 |
0.2660 |
Oaqa5b-tfidf |
0.6154 |
0.1429 |
0.2286 |
0.1810 |
0.1280 |
0.6109 |
0.1970 |
SemanticRoleLabeling |
0.4615 |
- | - | - |
0.2591 |
0.3962 |
0.3051 |
Lab Zhu,Fudan Univer |
0.4615 |
0.4000 |
0.5143 |
0.4524 |
0.3170 |
0.5801 |
0.3811 |
Basic QA pipline |
0.4615 |
0.1429 |
0.2571 |
0.1833 |
0.1818 |
0.2500 |
0.2062 |
LabZhu,FDU |
0.4615 |
0.4000 |
0.5143 |
0.4524 |
0.3619 |
0.6003 |
0.4188 |
LabZhu_FDU |
0.4615 |
- | - | - |
0.0091 |
0.0114 |
0.0101 |
LabZhu-FDU |
0.4615 |
0.1429 |
0.2000 |
0.1581 |
0.2146 |
0.2395 |
0.1783 |
UNCC System 2 |
0.4615 |
0.0571 |
0.1429 |
0.0786 |
0.0391 |
0.2867 |
0.0630 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
auth-qa-1 |
0.2629 |
0.2536 |
0.2735 |
0.2585 |
Olelo |
0.3417 |
0.1842 |
0.3518 |
0.1829 |
Olelo-GS |
0.2058 |
0.0544 |
0.2602 |
0.0646 |
auth-qa-2 |
- | - |
Deep QA (single) |
- | - |
Deep QA (ensemble) |
- | - |
BioASQ_Baseline |
- | - |
Lab Zhu ,Fdan Univer |
0.3236 |
0.3115 |
0.3232 |
0.3065 |
limsi-reader-UMLS-r2 |
- | - |
sarrouti |
0.5616 |
0.3706 |
0.5595 |
0.3601 |
MQ-1 |
0.5184 |
0.3587 |
0.5189 |
0.3500 |
MQ-2 |
0.5802 |
0.3982 |
0.5703 |
0.3813 |
MQ-3 |
0.4972 |
0.2883 |
0.4933 |
0.2784 |
MQ-4 |
0.4010 |
0.2180 |
0.4115 |
0.2165 |
Oaqa-5b |
0.6885 |
0.3613 |
0.6803 |
0.3481 |
Oaqa5b |
0.7064 |
0.3735 |
0.6962 |
0.3589 |
Oaqa 5b |
0.6771 |
0.3636 |
0.6672 |
0.3489 |
oaqa5b5 |
0.6771 |
0.3636 |
0.6672 |
0.3489 |
Oaqa5b-tfidf |
0.5773 |
0.3610 |
0.5747 |
0.3499 |
SemanticRoleLabeling |
0.0405 |
0.0402 |
0.0403 |
0.0393 |
Lab Zhu,Fudan Univer |
0.3369 |
0.3199 |
0.3358 |
0.3140 |
Basic QA pipline |
0.0697 |
0.0645 |
0.0705 |
0.0648 |
LabZhu,FDU |
0.3369 |
0.3199 |
0.3358 |
0.3140 |
LabZhu_FDU |
0.3236 |
0.3145 |
0.3232 |
0.3097 |
LabZhu-FDU |
0.3236 |
0.3145 |
0.3232 |
0.3097 |
UNCC System 2 |
0.7188 |
0.3509 |
0.7062 |
0.3354 |
+ Task 6b, Test batch 1
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
auth-qa-2 |
0.6875 |
0.1935 |
0.3548 |
0.2484 |
0.1545 |
0.5644 |
0.2320 |
auth-qa-1 |
0.6875 |
0.1935 |
0.3226 |
0.2376 |
0.1727 |
0.6023 |
0.2563 |
auth-qa-3 |
0.6875 |
0.1935 |
0.3226 |
0.2376 |
0.1636 |
0.5871 |
0.2450 |
Oaqa5b-tfidf |
- |
- | - | - |
- | - | - |
Oaqa5b |
- |
- | - | - |
- | - | - |
Oaqa 5b |
- |
- | - | - |
- | - | - |
Lab Zhu,Fudan Univer |
0.6875 |
0.1935 |
0.2258 |
0.2097 |
0.0904 |
0.4091 |
0.1413 |
LabZhu,FDU |
0.6875 |
0.1935 |
0.2258 |
0.2097 |
0.0934 |
0.4091 |
0.1459 |
MQ-1 |
0.6875 |
- | - | - |
- | - | - |
MQ-2 |
0.6875 |
- | - | - |
- | - | - |
MQ-3 |
0.6875 |
- | - | - |
- | - | - |
MQ-4 |
0.6875 |
- | - | - |
- | - | - |
MQ-5 |
0.6875 |
- | - | - |
- | - | - |
Oaqa-5b |
- |
- | - | - |
- | - | - |
OAQA based system |
0.5938 |
0.1613 |
0.3226 |
0.2366 |
0.3231 |
0.6023 |
0.4105 |
YODAQA based system |
0.6875 |
- | - | - |
0.0909 |
0.0909 |
0.0909 |
Lab Zhu ,Fdan Univer |
0.6875 |
0.2258 |
0.2581 |
0.2419 |
0.1015 |
0.4091 |
0.1580 |
LabZhu_FDU |
0.6875 |
0.0645 |
0.1290 |
0.0968 |
0.0707 |
0.4091 |
0.1165 |
LabZhu-FDU |
0.6875 |
0.0645 |
0.1290 |
0.0968 |
0.0707 |
0.4091 |
0.1165 |
SpanBaseline |
0.6875 |
0.0968 |
0.0968 |
0.0968 |
0.1818 |
0.0455 |
0.0714 |
oaqa5b5 |
- |
- | - | - |
- | - | - |
BioASQ_Baseline |
0.3125 |
0.2258 |
0.2903 |
0.2403 |
0.2463 |
0.4318 |
0.2807 |
Zero-shot BioBERT |
- |
0.2903 |
0.6452 |
0.4403 |
0.1455 |
0.2841 |
0.1843 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
auth-qa-2 |
- | - |
auth-qa-1 |
- | - |
auth-qa-3 |
- | - |
Oaqa5b-tfidf |
0.6621 |
0.3131 |
0.6651 |
0.3048 |
Oaqa5b |
0.5300 |
0.3171 |
0.5301 |
0.3101 |
Oaqa 5b |
0.6002 |
0.2854 |
0.5932 |
0.2725 |
Lab Zhu,Fudan Univer |
0.2742 |
0.2633 |
0.2776 |
0.2616 |
LabZhu,FDU |
0.2742 |
0.2633 |
0.2776 |
0.2616 |
MQ-1 |
0.4559 |
0.3095 |
0.4625 |
0.3049 |
MQ-2 |
0.5229 |
0.3529 |
0.5236 |
0.3425 |
MQ-3 |
0.4867 |
0.3267 |
0.4955 |
0.3219 |
MQ-4 |
0.5392 |
0.3545 |
0.5462 |
0.3487 |
MQ-5 |
0.3737 |
0.2823 |
0.3835 |
0.2813 |
Oaqa-5b |
0.6853 |
0.3281 |
0.6860 |
0.3176 |
OAQA based system |
0.0715 |
0.0928 |
0.0701 |
0.0911 |
YODAQA based system |
0.0118 |
0.0202 |
0.0087 |
0.0147 |
Lab Zhu ,Fdan Univer |
0.2742 |
0.2633 |
0.2776 |
0.2616 |
LabZhu_FDU |
0.2742 |
0.2633 |
0.2776 |
0.2616 |
LabZhu-FDU |
0.2742 |
0.2633 |
0.2776 |
0.2616 |
SpanBaseline |
0.0960 |
0.0982 |
0.0903 |
0.0875 |
oaqa5b5 |
0.6456 |
0.3022 |
0.6444 |
0.2898 |
BioASQ_Baseline |
- | - |
Zero-shot BioBERT |
- | - |
+ Task 6b, Test batch 2
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
auth-qa-1 |
0.7692 |
0.2381 |
0.5238 |
0.3468 |
0.1222 |
0.3271 |
0.1654 |
UNCC System 1 |
- |
- | - | - |
- | - | - |
auth-qa-2 |
0.7692 |
0.2381 |
0.6190 |
0.3746 |
0.1753 |
0.2993 |
0.1947 |
auth-qa-4 |
0.7692 |
0.2381 |
0.5714 |
0.3429 |
0.1966 |
0.3289 |
0.2252 |
auth-qa-5 |
0.7692 |
0.1905 |
0.5238 |
0.3032 |
0.1644 |
0.2623 |
0.1781 |
auth-qa-3 |
0.7692 |
0.2381 |
0.5238 |
0.3270 |
0.1389 |
0.3474 |
0.1834 |
limsi-reader-UMLS-r1 |
0.7692 |
0.2381 |
0.2857 |
0.2619 |
0.1111 |
0.1667 |
0.1270 |
MQ-1 |
0.7692 |
- | - | - |
- | - | - |
MQ-2 |
0.7692 |
- | - | - |
- | - | - |
MQ-3 |
0.7692 |
- | - | - |
- | - | - |
MQ-5 |
0.7692 |
- | - | - |
- | - | - |
MQ-4 |
0.7692 |
- | - | - |
- | - | - |
Oaqa5b-tfidf |
- |
- | - | - |
- | - | - |
Oaqa5b |
- |
- | - | - |
- | - | - |
Oaqa 5b |
- |
- | - | - |
- | - | - |
Oaqa-5b |
- |
- | - | - |
- | - | - |
Lab Zhu ,Fdan Univer |
0.7692 |
0.1905 |
0.3810 |
0.2619 |
0.0637 |
0.2789 |
0.0967 |
LabZhu-FDU |
0.7692 |
- | - | - |
- | - | - |
Lab Zhu,Fudan Univer |
0.7692 |
0.1905 |
0.3810 |
0.2619 |
0.1107 |
0.2743 |
0.1498 |
OAQA based system |
0.6923 |
0.2857 |
0.2857 |
0.2857 |
0.2086 |
0.2843 |
0.2109 |
LabZhu,FDU |
0.7692 |
0.3333 |
0.5714 |
0.4325 |
0.2044 |
0.2761 |
0.2279 |
YODAQA based system |
0.7692 |
0.0952 |
0.1905 |
0.1429 |
0.1944 |
0.0917 |
0.1168 |
BioASQ_Baseline |
0.3846 |
0.1429 |
0.2857 |
0.1841 |
0.1810 |
0.2689 |
0.2107 |
Zero-shot BioBERT |
- |
0.3810 |
0.6667 |
0.5040 |
0.0667 |
0.1074 |
0.0788 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
auth-qa-1 |
- | - |
UNCC System 1 |
0.6090 |
0.3051 |
0.6176 |
0.2981 |
auth-qa-2 |
- | - |
auth-qa-4 |
- | - |
auth-qa-5 |
- | - |
auth-qa-3 |
- | - |
limsi-reader-UMLS-r1 |
- | - |
MQ-1 |
0.4955 |
0.3246 |
0.5033 |
0.3200 |
MQ-2 |
0.4882 |
0.3197 |
0.4971 |
0.3147 |
MQ-3 |
0.4957 |
0.3278 |
0.5107 |
0.3257 |
MQ-5 |
0.4206 |
0.3134 |
0.4310 |
0.3128 |
MQ-4 |
0.5289 |
0.3467 |
0.5385 |
0.3397 |
Oaqa5b-tfidf |
0.4842 |
0.3022 |
0.4943 |
0.2982 |
Oaqa5b |
0.5327 |
0.2656 |
0.5311 |
0.2575 |
Oaqa 5b |
0.5704 |
0.2984 |
0.5792 |
0.2929 |
Oaqa-5b |
0.5526 |
0.2910 |
0.5581 |
0.2842 |
Lab Zhu ,Fdan Univer |
0.3103 |
0.2925 |
0.3202 |
0.2895 |
LabZhu-FDU |
0.3103 |
0.2925 |
0.3202 |
0.2895 |
Lab Zhu,Fudan Univer |
0.3103 |
0.2925 |
0.3202 |
0.2895 |
OAQA based system |
0.1257 |
0.1313 |
0.1272 |
0.1314 |
LabZhu,FDU |
0.3103 |
0.2925 |
0.3202 |
0.2895 |
YODAQA based system |
0.0273 |
0.0340 |
0.0245 |
0.0294 |
BioASQ_Baseline |
- | - |
Zero-shot BioBERT |
- | - |
+ Task 6b, Test batch 3
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
auth-qa-1 |
0.6800 |
0.1563 |
0.4063 |
0.2604 |
0.1462 |
0.4013 |
0.2068 |
auth-qa-2 |
0.6800 |
0.1563 |
0.3125 |
0.2135 |
0.2120 |
0.3244 |
0.2333 |
auth-qa-3 |
0.6800 |
0.2500 |
0.4375 |
0.3083 |
0.1538 |
0.4577 |
0.2238 |
auth-qa-4 |
0.6800 |
0.2188 |
0.3750 |
0.2656 |
0.1982 |
0.3628 |
0.2311 |
UNCC System 1 |
- |
- | - | - |
- | - | - |
UNCC System 2 |
- |
- | - | - |
- | - | - |
MQ-1 |
0.6800 |
- | - | - |
- | - | - |
MQ-2 |
0.6800 |
- | - | - |
- | - | - |
MQ-3 |
0.6800 |
- | - | - |
- | - | - |
MQ-4 |
0.6800 |
- | - | - |
- | - | - |
MQ-5 |
0.6800 |
- | - | - |
- | - | - |
YODAQA based system |
0.6800 |
0.0625 |
0.1563 |
0.1094 |
0.1154 |
0.0833 |
0.0949 |
Oaqa5b-tfidf |
- |
- | - | - |
- | - | - |
Oaqa5b |
- |
- | - | - |
- | - | - |
OAQA based system |
0.6400 |
0.1250 |
0.3438 |
0.2094 |
0.1973 |
0.3538 |
0.2432 |
Oaqa-5b |
- |
- | - | - |
- | - | - |
oaqa5b5 |
- |
- | - | - |
- | - | - |
Lab Zhu ,Fdan Univer |
0.6800 |
0.0938 |
0.2500 |
0.1589 |
0.1094 |
0.3654 |
0.1502 |
Lab Zhu,Fudan Univer |
0.6800 |
0.1875 |
0.3125 |
0.2370 |
0.2675 |
0.4103 |
0.3051 |
LabZhu,FDU |
0.6800 |
0.1875 |
0.3125 |
0.2370 |
0.2803 |
0.4103 |
0.3216 |
BioASQ_Baseline |
0.5200 |
0.2188 |
0.2813 |
0.2396 |
0.1406 |
0.3923 |
0.1859 |
Zero-shot BioBERT |
- |
0.4688 |
0.6250 |
0.5339 |
0.0923 |
0.1231 |
0.1006 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
auth-qa-1 |
- | - |
auth-qa-2 |
- | - |
auth-qa-3 |
- | - |
auth-qa-4 |
- | - |
UNCC System 1 |
0.6445 |
0.3128 |
0.6522 |
0.3039 |
UNCC System 2 |
0.2324 |
0.2396 |
0.2300 |
0.2359 |
MQ-1 |
0.4467 |
0.3122 |
0.4468 |
0.3069 |
MQ-2 |
0.4461 |
0.3046 |
0.4508 |
0.3002 |
MQ-3 |
0.4856 |
0.3289 |
0.4915 |
0.3248 |
MQ-4 |
0.5041 |
0.3443 |
0.5112 |
0.3396 |
MQ-5 |
0.3782 |
0.2850 |
0.3820 |
0.2816 |
YODAQA based system |
0.0143 |
0.0222 |
0.0113 |
0.0175 |
Oaqa5b-tfidf |
0.6016 |
0.3214 |
0.6129 |
0.3166 |
Oaqa5b |
0.4746 |
0.3002 |
0.4872 |
0.2963 |
OAQA based system |
0.1431 |
0.1345 |
0.1430 |
0.1335 |
Oaqa-5b |
0.5739 |
0.3027 |
0.5741 |
0.2924 |
oaqa5b5 |
0.5454 |
0.3083 |
0.5479 |
0.2995 |
Lab Zhu ,Fdan Univer |
0.2907 |
0.2583 |
0.2908 |
0.2533 |
Lab Zhu,Fudan Univer |
0.2941 |
0.2614 |
0.2948 |
0.2570 |
LabZhu,FDU |
0.2941 |
0.2614 |
0.2948 |
0.2570 |
BioASQ_Baseline |
- | - |
Zero-shot BioBERT |
- | - |
+ Task 6b, Test batch 4
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
auth-qa-2 |
0.6296 |
0.2121 |
0.3030 |
0.2475 |
0.2511 |
0.3822 |
0.2925 |
auth-qa-4 |
0.6296 |
0.2121 |
0.3030 |
0.2434 |
0.2800 |
0.3822 |
0.3094 |
auth-qa-1 |
0.6296 |
0.2121 |
0.2727 |
0.2374 |
0.1600 |
0.4267 |
0.2277 |
auth-qa-3 |
0.6296 |
0.2121 |
0.2727 |
0.2283 |
0.1800 |
0.4711 |
0.2551 |
Lab Zhu ,Fdan Univer |
0.6296 |
0.0909 |
0.1212 |
0.1061 |
0.1657 |
0.2833 |
0.1663 |
Oaqa5b-tfidf |
- |
- | - | - |
- | - | - |
Oaqa5b |
- |
- | - | - |
- | - | - |
Oaqa 5b |
- |
- | - | - |
- | - | - |
Oaqa-5b |
0.6667 |
0.0606 |
0.2121 |
0.1313 |
0.0867 |
0.2722 |
0.1299 |
MQ-1 |
0.6296 |
- | - | - |
- | - | - |
MQ-2 |
0.6296 |
- | - | - |
- | - | - |
MQ-3 |
0.6296 |
- | - | - |
- | - | - |
MQ-4 |
0.6296 |
- | - | - |
- | - | - |
MQ-5 |
0.6296 |
- | - | - |
- | - | - |
auth-qa-5 |
0.6296 |
0.2121 |
0.3030 |
0.2434 |
0.2800 |
0.3822 |
0.3094 |
Lab Zhu,Fudan Univer |
0.6296 |
0.2121 |
0.2424 |
0.2273 |
0.2944 |
0.3411 |
0.2902 |
LabZhu,FDU |
0.6296 |
0.2424 |
0.2424 |
0.2424 |
0.4130 |
0.3356 |
0.3280 |
BioASQ_Baseline |
0.4815 |
0.0606 |
0.1212 |
0.0859 |
0.1774 |
0.3811 |
0.2227 |
Zero-shot BioBERT |
- |
0.3636 |
0.4848 |
0.4101 |
0.1333 |
0.2189 |
0.1521 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
auth-qa-2 |
- | - |
auth-qa-4 |
- | - |
auth-qa-1 |
- | - |
auth-qa-3 |
- | - |
Lab Zhu ,Fdan Univer |
0.3004 |
0.2643 |
0.3095 |
0.2612 |
Oaqa5b-tfidf |
0.5390 |
0.3144 |
0.5425 |
0.3039 |
Oaqa5b |
0.5477 |
0.3153 |
0.5513 |
0.3045 |
Oaqa 5b |
0.4311 |
0.3128 |
0.4421 |
0.3090 |
Oaqa-5b |
0.6292 |
0.3256 |
0.6398 |
0.3184 |
MQ-1 |
0.4967 |
0.3348 |
0.5101 |
0.3306 |
MQ-2 |
0.5291 |
0.3507 |
0.5307 |
0.3410 |
MQ-3 |
0.5103 |
0.3316 |
0.5160 |
0.3240 |
MQ-4 |
0.5578 |
0.3619 |
0.5653 |
0.3531 |
MQ-5 |
0.4473 |
0.3393 |
0.4624 |
0.3363 |
auth-qa-5 |
- | - |
Lab Zhu,Fudan Univer |
0.3004 |
0.2643 |
0.3095 |
0.2612 |
LabZhu,FDU |
0.3004 |
0.2643 |
0.3095 |
0.2612 |
BioASQ_Baseline |
- | - |
Zero-shot BioBERT |
- | - |
+ Task 6b, Test batch 5
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
Lab Zhu ,Fdan Univer |
0.7000 |
0.0909 |
0.1818 |
0.1250 |
0.0960 |
0.1808 |
0.1141 |
auth-qa-2 |
0.7000 |
0.1136 |
0.2727 |
0.1758 |
0.2143 |
0.2580 |
0.2187 |
auth-qa-4 |
0.7000 |
0.1136 |
0.2727 |
0.1758 |
0.1952 |
0.2419 |
0.1949 |
auth-qa-1 |
0.7000 |
0.1136 |
0.2955 |
0.1758 |
0.1500 |
0.3858 |
0.2070 |
auth-qa-3 |
0.7000 |
0.1136 |
0.2955 |
0.1777 |
0.1429 |
0.3654 |
0.1980 |
UNCC System 1 |
- |
- | - | - |
- | - | - |
UNCC System 2 |
- |
- | - | - |
- | - | - |
Oaqa5b-tfidf |
- |
- | - | - |
- | - | - |
Oaqa5b |
- |
- | - | - |
- | - | - |
Oaqa 5b |
- |
- | - | - |
- | - | - |
MQ-1 |
0.7000 |
- | - | - |
- | - | - |
MQ-2 |
0.7000 |
- | - | - |
- | - | - |
MQ-3 |
0.7000 |
- | - | - |
- | - | - |
MQ-4 |
0.7000 |
- | - | - |
- | - | - |
MQ-5 |
0.7000 |
- | - | - |
- | - | - |
Lab Zhu,Fudan Univer |
0.7000 |
0.1136 |
0.2045 |
0.1477 |
0.0960 |
0.1808 |
0.1141 |
Olelo system at HPI |
0.7000 |
- | - | - |
0.0204 |
0.0204 |
0.0204 |
LabZhu,FDU |
0.7000 |
0.1136 |
0.2045 |
0.1477 |
0.0960 |
0.1808 |
0.1141 |
LabZhu_FDU |
- |
0.0455 |
0.0682 |
0.0511 |
0.0560 |
0.0316 |
0.0397 |
auth-qa-5 |
0.4500 |
0.1136 |
0.2727 |
0.1758 |
0.1952 |
0.2419 |
0.1949 |
Oaqa-5b |
- |
- | - | - |
- | - | - |
oaqa5b5 |
0.6000 |
0.1818 |
0.2273 |
0.1951 |
0.0929 |
0.1757 |
0.1135 |
LabZhu-FDU |
0.7000 |
0.2045 |
0.2727 |
0.2273 |
0.1736 |
0.2344 |
0.1832 |
BioASQ_Baseline |
0.6000 |
- | - | - |
0.2862 |
0.2999 |
0.2544 |
Zero-shot BioBERT |
- |
0.3409 |
0.5682 |
0.4364 |
0.1429 |
0.1904 |
0.1506 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
Lab Zhu ,Fdan Univer |
0.4334 |
0.3909 |
0.4273 |
0.3828 |
auth-qa-2 |
- | - |
auth-qa-4 |
- | - |
auth-qa-1 |
- | - |
auth-qa-3 |
- | - |
UNCC System 1 |
0.7607 |
0.3788 |
0.7483 |
0.3592 |
UNCC System 2 |
0.4329 |
0.4215 |
0.4261 |
0.4130 |
Oaqa5b-tfidf |
0.6427 |
0.3838 |
0.6386 |
0.3684 |
Oaqa5b |
0.6782 |
0.3583 |
0.6657 |
0.3388 |
Oaqa 5b |
0.6762 |
0.3659 |
0.6663 |
0.3467 |
MQ-1 |
0.5936 |
0.4050 |
0.5890 |
0.3920 |
MQ-2 |
0.5942 |
0.3995 |
0.5869 |
0.3828 |
MQ-3 |
0.6207 |
0.4148 |
0.6167 |
0.4013 |
MQ-4 |
0.6439 |
0.4225 |
0.6414 |
0.4092 |
MQ-5 |
0.5220 |
0.4032 |
0.5196 |
0.3929 |
Lab Zhu,Fudan Univer |
0.4334 |
0.3909 |
0.4273 |
0.3828 |
Olelo system at HPI |
0.1880 |
0.0495 |
0.2352 |
0.0606 |
LabZhu,FDU |
0.4334 |
0.3909 |
0.4273 |
0.3828 |
LabZhu_FDU |
0.4334 |
0.3909 |
0.4273 |
0.3828 |
auth-qa-5 |
- | - |
Oaqa-5b |
0.4854 |
0.3128 |
0.4901 |
0.3056 |
oaqa5b5 |
0.2588 |
0.2747 |
0.2538 |
0.2674 |
LabZhu-FDU |
0.4334 |
0.3909 |
0.4273 |
0.3828 |
BioASQ_Baseline |
- | - |
Zero-shot BioBERT |
- | - |
+ Task 6b, Test batch 6
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
transfer-learning |
0.7931 |
0.1795 |
0.2051 |
0.1880 |
- | - | - |
Lab Zhu ,Fdan Univer |
0.7931 |
0.1538 |
0.2308 |
0.1923 |
0.2347 |
0.5748 |
0.3165 |
auth-qa-1 |
0.7586 |
0.2564 |
0.3077 |
0.2778 |
0.1833 |
0.4877 |
0.2594 |
MQ-3 |
0.7931 |
- | - | - |
- | - | - |
MQ-4 |
0.7931 |
- | - | - |
- | - | - |
MQ-1 |
0.7931 |
- | - | - |
- | - | - |
MQ-2 |
0.7931 |
- | - | - |
- | - | - |
MQ-5 |
0.7931 |
- | - | - |
- | - | - |
QA1 |
0.7931 |
0.1538 |
0.2308 |
0.1761 |
- | - | - |
Lab Zhu,Fudan Univer |
0.7931 |
0.2308 |
0.3077 |
0.2692 |
0.2417 |
0.6026 |
0.3276 |
LabZhu,FDU |
0.7931 |
0.0256 |
0.1026 |
0.0577 |
0.1424 |
0.4855 |
0.2153 |
KU-DMIS-5 |
0.7931 |
- | - | - |
- | - | - |
LabZhu-FDU |
0.7931 |
0.0513 |
0.0769 |
0.0641 |
0.0417 |
0.1452 |
0.0620 |
LabZhu_FDU |
0.7931 |
0.0769 |
0.1282 |
0.0962 |
0.1215 |
0.4230 |
0.1840 |
KU-DMIS-1 |
0.8276 |
0.4103 |
0.5385 |
0.4637 |
0.4792 |
0.2476 |
0.3051 |
BJUTNLPGroup |
- |
0.3077 |
0.4103 |
0.3483 |
0.1500 |
0.2411 |
0.1785 |
BioASQ_Baseline |
0.4828 |
0.1282 |
0.2051 |
0.1547 |
0.2266 |
0.3460 |
0.2444 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
transfer-learning |
- | - |
Lab Zhu ,Fdan Univer |
0.2964 |
0.2700 |
0.2968 |
0.2638 |
auth-qa-1 |
- | - |
MQ-3 |
0.5008 |
0.3467 |
0.4973 |
0.3343 |
MQ-4 |
0.5467 |
0.3771 |
0.5443 |
0.3646 |
MQ-1 |
0.5008 |
0.3467 |
0.4973 |
0.3343 |
MQ-2 |
0.5469 |
0.3785 |
0.5370 |
0.3619 |
MQ-5 |
0.4806 |
0.3551 |
0.4838 |
0.3479 |
QA1 |
- | - |
Lab Zhu,Fudan Univer |
0.2964 |
0.2700 |
0.2968 |
0.2638 |
LabZhu,FDU |
0.2964 |
0.2700 |
0.2968 |
0.2638 |
KU-DMIS-5 |
- | - |
LabZhu-FDU |
0.2964 |
0.2700 |
0.2968 |
0.2638 |
LabZhu_FDU |
0.2964 |
0.2700 |
0.2968 |
0.2638 |
KU-DMIS-1 |
- | - |
BJUTNLPGroup |
- | - |
BioASQ_Baseline |
- | - |
+ Task 6b, Test batch 7
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
auth-qa-1 |
0.6333 |
0.2400 |
0.4000 |
0.3067 |
0.1549 |
0.5143 |
0.2321 |
auth-qa-2 |
0.5000 |
0.2400 |
0.4000 |
0.3067 |
0.1549 |
0.5143 |
0.2321 |
transfer-learning |
0.5667 |
0.2400 |
0.4400 |
0.3267 |
- | - | - |
limsi-reader |
0.5667 |
0.1600 |
0.3200 |
0.2200 |
- | - | - |
Lab Zhu ,Fdan Univer |
0.5667 |
0.0800 |
0.2000 |
0.1300 |
0.1232 |
0.4422 |
0.1865 |
limsi-reader-UMLS-r1 |
0.5667 |
0.1600 |
0.3200 |
0.2100 |
- | - | - |
Lab Zhu,Fudan Univer |
0.5667 |
0.1600 |
0.2800 |
0.2100 |
0.1732 |
0.4814 |
0.2450 |
LabZhu,FDU |
0.5667 |
0.2000 |
0.3200 |
0.2500 |
0.1818 |
0.5108 |
0.2579 |
LabZhu_FDU |
0.5667 |
0.0400 |
0.1200 |
0.0800 |
0.1103 |
0.1431 |
0.0998 |
MQ-1 |
0.5667 |
- | - | - |
- | - | - |
MQ-2 |
0.5667 |
- | - | - |
- | - | - |
QA1 |
0.5667 |
0.3600 |
0.4800 |
0.4033 |
0.0471 |
0.2898 |
0.0786 |
MQ-3 |
0.5667 |
- | - | - |
- | - | - |
MQ-4 |
0.5667 |
- | - | - |
- | - | - |
MQ-5 |
0.5667 |
- | - | - |
- | - | - |
LabZhu-FDU |
0.5667 |
0.0400 |
0.1600 |
0.0900 |
0.1103 |
0.1431 |
0.0998 |
List only |
0.5667 |
0.2400 |
0.3200 |
0.2733 |
- | - | - |
L2PS - DeepQA |
0.5667 |
0.1600 |
0.3200 |
0.2333 |
- | - | - |
KU-DMIS-1 |
0.5667 |
0.3200 |
0.6000 |
0.4367 |
0.5826 |
0.4839 |
0.4732 |
KU-DMIS-5 |
0.8333 |
0.5200 |
0.6400 |
0.5667 |
0.5696 |
0.4368 |
0.4395 |
BioASQ_Baseline |
0.4667 |
0.0800 |
0.2400 |
0.1367 |
0.1687 |
0.2954 |
0.1823 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
auth-qa-1 |
- | - |
auth-qa-2 |
- | - |
transfer-learning |
- | - |
limsi-reader |
- | - |
Lab Zhu ,Fdan Univer |
0.3720 |
0.3334 |
0.3771 |
0.3293 |
limsi-reader-UMLS-r1 |
- | - |
Lab Zhu,Fudan Univer |
0.3720 |
0.3334 |
0.3771 |
0.3293 |
LabZhu,FDU |
0.3720 |
0.3334 |
0.3771 |
0.3293 |
LabZhu_FDU |
0.3720 |
0.3334 |
0.3771 |
0.3293 |
MQ-1 |
0.5102 |
0.3496 |
0.5251 |
0.3453 |
MQ-2 |
0.5120 |
0.3443 |
0.5285 |
0.3402 |
QA1 |
- | - |
MQ-3 |
0.5102 |
0.3496 |
0.5251 |
0.3453 |
MQ-4 |
0.5279 |
0.3565 |
0.5456 |
0.3524 |
MQ-5 |
0.4795 |
0.3438 |
0.4989 |
0.3413 |
LabZhu-FDU |
0.3720 |
0.3334 |
0.3771 |
0.3293 |
List only |
- | - |
L2PS - DeepQA |
- | - |
KU-DMIS-1 |
- | - |
KU-DMIS-5 |
- | - |
BioASQ_Baseline |
- | - |
+ Task 6b, Test batch 8
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
auth-qa-1 |
0.5652 |
0.2759 |
0.4483 |
0.3420 |
0.1720 |
0.5338 |
0.2513 |
auth-qa-2 |
0.6957 |
0.2759 |
0.4483 |
0.3420 |
0.1720 |
0.5338 |
0.2513 |
google-gold-input |
0.7826 |
0.4138 |
0.6552 |
0.5023 |
- | - | - |
google-pred-input |
0.7826 |
0.3448 |
0.5517 |
0.4322 |
- | - | - |
Lab Zhu ,Fdan Univer |
0.7826 |
0.1379 |
0.3103 |
0.2023 |
0.1117 |
0.2689 |
0.1381 |
unipi-quokka-QA-1 |
0.8261 |
- | - | - |
- | - | - |
QA1 |
0.7826 |
0.4483 |
0.5862 |
0.5115 |
0.0780 |
0.4711 |
0.1297 |
UNCC_QA_1 |
0.7826 |
0.4483 |
0.5862 |
0.5115 |
0.0780 |
0.4711 |
0.1297 |
unipi-quokka-QA-2 |
0.8696 |
- | - | - |
- | - | - |
MQ-1 |
0.7826 |
- | - | - |
- | - | - |
MQ-2 |
0.7826 |
- | - | - |
- | - | - |
MQ-3 |
0.7826 |
- | - | - |
- | - | - |
MQ-4 |
0.7826 |
- | - | - |
- | - | - |
MQ-5 |
0.7826 |
- | - | - |
- | - | - |
UNCC_QA2 |
0.7826 |
0.4138 |
0.5862 |
0.4856 |
0.0680 |
0.3913 |
0.1121 |
UNCC_QA3 |
0.7826 |
0.4138 |
0.5862 |
0.4943 |
0.0780 |
0.4711 |
0.1297 |
limsi-reader |
0.7826 |
0.2414 |
0.4828 |
0.3213 |
- | - | - |
limsi-reader-UMLS-r1 |
0.7826 |
0.2414 |
0.4828 |
0.3126 |
- | - | - |
KU-DMIS-1 |
0.6087 |
0.3793 |
0.6207 |
0.4724 |
0.4267 |
0.3058 |
0.3298 |
Lab Zhu,Fudan Univer |
0.7826 |
0.2759 |
0.4828 |
0.3621 |
0.1490 |
0.3489 |
0.1875 |
LabZhu,FDU |
0.7826 |
0.3103 |
0.5172 |
0.3966 |
0.1847 |
0.3822 |
0.2261 |
LabZhu_FDU |
0.7826 |
0.1379 |
0.2069 |
0.1667 |
0.0967 |
0.1649 |
0.1114 |
BJUTNLPGroup |
0.7826 |
0.2759 |
0.3793 |
0.3011 |
0.0960 |
0.1791 |
0.1201 |
BioASQ_Baseline |
0.1739 |
0.1034 |
0.1724 |
0.1322 |
0.1928 |
0.4080 |
0.2275 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
auth-qa-1 |
- | - |
auth-qa-2 |
- | - |
google-gold-input |
- | - |
google-pred-input |
- | - |
Lab Zhu ,Fdan Univer |
0.2793 |
0.2577 |
0.2827 |
0.2575 |
unipi-quokka-QA-1 |
0.0525 |
0.0466 |
0.0566 |
0.0489 |
QA1 |
- | - |
UNCC_QA_1 |
- | - |
unipi-quokka-QA-2 |
0.0525 |
0.0466 |
0.0566 |
0.0489 |
MQ-1 |
0.4790 |
0.2897 |
0.4850 |
0.2833 |
MQ-2 |
0.5013 |
0.3001 |
0.5086 |
0.2950 |
MQ-3 |
0.4790 |
0.2897 |
0.4850 |
0.2833 |
MQ-4 |
0.5309 |
0.3169 |
0.5344 |
0.3088 |
MQ-5 |
0.4667 |
0.3144 |
0.4737 |
0.3095 |
UNCC_QA2 |
- | - |
UNCC_QA3 |
- | - |
limsi-reader |
- | - |
limsi-reader-UMLS-r1 |
- | - |
KU-DMIS-1 |
- | - |
Lab Zhu,Fudan Univer |
0.2793 |
0.2577 |
0.2827 |
0.2575 |
LabZhu,FDU |
0.2793 |
0.2577 |
0.2827 |
0.2575 |
LabZhu_FDU |
0.2793 |
0.2577 |
0.2827 |
0.2575 |
BJUTNLPGroup |
0.0942 |
0.1296 |
0.0781 |
0.1049 |
BioASQ_Baseline |
- | - |
+ Task 6b, Test batch 9
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
auth-qa-1 |
0.5217 |
0.3235 |
0.4706 |
0.3725 |
0.1864 |
0.5302 |
0.2652 |
auth-qa-2 |
0.7391 |
0.3235 |
0.4706 |
0.3725 |
0.1864 |
0.5302 |
0.2652 |
MQ-1 |
0.7391 |
- | - | - |
- | - | - |
MQ-2 |
0.7391 |
- | - | - |
- | - | - |
MQ-3 |
0.7391 |
- | - | - |
- | - | - |
MQ-4 |
0.7391 |
- | - | - |
- | - | - |
MQ-5 |
0.7391 |
- | - | - |
- | - | - |
auth-qa-3 |
0.6522 |
0.3235 |
0.4706 |
0.3725 |
0.1864 |
0.5302 |
0.2652 |
limsi-reader |
0.7391 |
0.2941 |
0.5882 |
0.4054 |
- | - | - |
google-gold-input-ab |
0.7391 |
0.4706 |
0.6471 |
0.5255 |
0.3236 |
0.5773 |
0.4001 |
google-pred-input |
0.7391 |
0.3529 |
0.5882 |
0.4338 |
0.1385 |
0.2966 |
0.1806 |
google-gold-input-nq |
0.7391 |
0.4706 |
0.5882 |
0.5132 |
0.3562 |
0.6000 |
0.4315 |
google-gold-input |
0.7391 |
0.4706 |
0.7059 |
0.5495 |
0.3246 |
0.5805 |
0.4003 |
limsi-reader-UMLS-r1 |
0.7391 |
0.2059 |
0.4412 |
0.2902 |
- | - | - |
Lab Zhu ,Fdan Univer |
0.7391 |
0.2353 |
0.3235 |
0.2706 |
0.1098 |
0.3540 |
0.1623 |
UNCC_QA_1 |
0.6087 |
0.4706 |
0.7353 |
0.5833 |
0.1087 |
0.6892 |
0.1843 |
BJUTNLPGroup |
0.7391 |
0.2353 |
0.6471 |
0.3745 |
0.2273 |
0.3781 |
0.2755 |
BJUTNLPGroup_v2 |
0.7391 |
0.2353 |
0.5000 |
0.3152 |
0.1000 |
0.1504 |
0.1169 |
FACTOIDS |
0.7391 |
0.5294 |
0.7353 |
0.6103 |
0.1119 |
0.6957 |
0.1890 |
UNCC_QA3 |
0.7391 |
0.5294 |
0.7353 |
0.6103 |
0.1087 |
0.6892 |
0.1843 |
L2PS - DeepQA |
0.7391 |
0.2353 |
0.4118 |
0.2760 |
- | - | - |
List only |
0.7391 |
0.0294 |
0.3235 |
0.1368 |
- | - | - |
Lab Zhu,Fudan Univer |
0.7391 |
0.3235 |
0.4412 |
0.3686 |
0.2130 |
0.4600 |
0.2683 |
LabZhu,FDU |
0.7391 |
0.4412 |
0.5588 |
0.4863 |
0.2753 |
0.4777 |
0.3192 |
limsi-reader-UMLS-r2 |
0.7391 |
0.2059 |
0.5000 |
0.3309 |
- | - | - |
KU-DMIS-1 |
0.7391 |
0.5882 |
0.8235 |
0.6912 |
0.4841 |
0.4937 |
0.4539 |
KU-DMIS-2 |
0.7391 |
0.5882 |
0.8235 |
0.6863 |
0.3828 |
0.4369 |
0.3769 |
KU-DMIS-3 |
0.8261 |
0.5588 |
0.7941 |
0.6593 |
0.5024 |
0.4141 |
0.4068 |
KU-DMIS-4 |
0.8696 |
0.4706 |
0.7353 |
0.5696 |
0.5024 |
0.4141 |
0.4068 |
unipi-quokka-QA-1 |
0.7391 |
0.2059 |
0.3824 |
0.2730 |
0.1442 |
0.6193 |
0.2163 |
unipi-quokka-QA-2 |
0.7391 |
0.2059 |
0.3824 |
0.2730 |
0.1442 |
0.6193 |
0.2163 |
unipi-quokka-QA-3 |
0.8261 |
0.2059 |
0.3824 |
0.2730 |
0.1442 |
0.6193 |
0.2163 |
unipi-quokka-QA-4 |
0.8696 |
0.2059 |
0.3824 |
0.2730 |
0.1442 |
0.6193 |
0.2163 |
KU-DMIS-5 |
0.7391 |
0.4706 |
0.7941 |
0.5990 |
0.3242 |
0.5813 |
0.3714 |
bioasq_experiments |
0.6087 |
0.0588 |
0.0882 |
0.0735 |
0.0455 |
0.0303 |
0.0364 |
BioASQ_Baseline |
0.4348 |
0.1765 |
0.4118 |
0.2534 |
0.1878 |
0.4225 |
0.2398 |
BioASQ GrantID |
- |
- | - | - |
- | - | - |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
auth-qa-1 |
- | - |
auth-qa-2 |
- | - |
MQ-1 |
0.4867 |
0.3097 |
0.4976 |
0.3006 |
MQ-2 |
0.5050 |
0.3221 |
0.5159 |
0.3102 |
MQ-3 |
0.4867 |
0.3097 |
0.4976 |
0.3006 |
MQ-4 |
0.5537 |
0.3504 |
0.5618 |
0.3386 |
MQ-5 |
0.4239 |
0.2879 |
0.4340 |
0.2810 |
auth-qa-3 |
- | - |
limsi-reader |
- | - |
google-gold-input-ab |
- | - |
google-pred-input |
- | - |
google-gold-input-nq |
- | - |
google-gold-input |
- | - |
limsi-reader-UMLS-r1 |
- | - |
Lab Zhu ,Fdan Univer |
0.3081 |
0.2549 |
0.3241 |
0.2574 |
UNCC_QA_1 |
- | - |
BJUTNLPGroup |
0.0433 |
0.0630 |
0.0323 |
0.0476 |
BJUTNLPGroup_v2 |
0.0996 |
0.1264 |
0.0880 |
0.1098 |
FACTOIDS |
- | - |
UNCC_QA3 |
- | - |
L2PS - DeepQA |
- | - |
List only |
- | - |
Lab Zhu,Fudan Univer |
0.3081 |
0.2549 |
0.3241 |
0.2574 |
LabZhu,FDU |
0.3081 |
0.2549 |
0.3241 |
0.2574 |
limsi-reader-UMLS-r2 |
- | - |
KU-DMIS-1 |
- | - |
KU-DMIS-2 |
- | - |
KU-DMIS-3 |
- | - |
KU-DMIS-4 |
- | - |
unipi-quokka-QA-1 |
0.3872 |
0.2745 |
0.4012 |
0.2697 |
unipi-quokka-QA-2 |
0.3872 |
0.2745 |
0.4012 |
0.2697 |
unipi-quokka-QA-3 |
0.3872 |
0.2745 |
0.4012 |
0.2697 |
unipi-quokka-QA-4 |
0.3872 |
0.2745 |
0.4012 |
0.2697 |
KU-DMIS-5 |
- | - |
bioasq_experiments |
0.4094 |
0.2795 |
0.4240 |
0.2744 |
BioASQ_Baseline |
- | - |
BioASQ GrantID |
- | - |
+ Task 6b, Test batch 10
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
auth-qa-1 |
0.5714 |
0.0857 |
0.2286 |
0.1500 |
0.1361 |
0.4667 |
0.2075 |
auth-qa-2 |
0.5429 |
0.0857 |
0.2286 |
0.1500 |
0.1361 |
0.4667 |
0.2075 |
Neural MSE |
- |
- | - | - |
- | - | - |
First K sentences |
- |
- | - | - |
- | - | - |
auth-qa-3 |
0.6286 |
0.0857 |
0.2286 |
0.1500 |
0.1361 |
0.4667 |
0.2075 |
MQ-1 |
0.5429 |
- | - | - |
- | - | - |
MQ-2 |
0.5429 |
- | - | - |
- | - | - |
MQ-5 |
0.5429 |
- | - | - |
- | - | - |
MQ-3 |
0.5429 |
- | - | - |
- | - | - |
auth-qa-4 |
0.6286 |
0.0857 |
0.2286 |
0.1500 |
0.1361 |
0.4667 |
0.2075 |
MQ-4 |
0.5429 |
- | - | - |
- | - | - |
google-gold-input-ab |
0.7143 |
0.2286 |
0.2857 |
0.2571 |
0.1774 |
0.4175 |
0.2415 |
google-pred-input |
0.6286 |
0.1429 |
0.2857 |
0.2057 |
0.0863 |
0.2222 |
0.1222 |
google-gold-input-nq |
0.6286 |
0.2857 |
0.3714 |
0.3057 |
0.2218 |
0.4452 |
0.2889 |
google-gold-input |
0.6571 |
0.2857 |
0.3714 |
0.3167 |
0.2159 |
0.4452 |
0.2824 |
Lab Zhu ,Fdan Univer |
0.5429 |
0.0286 |
0.0571 |
0.0381 |
0.1736 |
0.2925 |
0.1932 |
System B process |
- |
- | - | - |
- | - | - |
BJUTNLPGroup |
0.5429 |
0.2857 |
0.4000 |
0.3381 |
0.1667 |
0.2813 |
0.2020 |
UNCC_QA_1 |
0.4857 |
0.2857 |
0.4286 |
0.3305 |
0.2051 |
0.5127 |
0.2862 |
UNCC_QA3 |
0.4857 |
0.2286 |
0.3143 |
0.2643 |
0.2051 |
0.5127 |
0.2862 |
UNCC_QA2 |
0.5429 |
0.2857 |
0.4286 |
0.3305 |
0.2051 |
0.5127 |
0.2862 |
System B+C proc |
- |
- | - | - |
- | - | - |
System D+E process |
- |
- | - | - |
- | - | - |
System384 B process |
- |
- | - | - |
- | - | - |
System384 D+E proces |
- |
- | - | - |
- | - | - |
limsi-reader |
0.5429 |
0.2000 |
0.4000 |
0.2748 |
- | - | - |
Neural MSE Attention |
- |
- | - | - |
- | - | - |
L2PS - DeepQA |
0.5429 |
0.0571 |
0.2571 |
0.1257 |
- | - | - |
List only |
0.5429 |
0.0857 |
0.2571 |
0.1352 |
- | - | - |
limsi-reader-UMLS-r2 |
0.5429 |
0.2286 |
0.4000 |
0.2890 |
- | - | - |
KU-DMIS-1 |
0.8000 |
0.2571 |
0.4571 |
0.3224 |
0.5236 |
0.3714 |
0.4202 |
Lab Zhu,Fudan Univer |
0.5429 |
0.1143 |
0.1714 |
0.1381 |
0.2278 |
0.3091 |
0.2536 |
KU-DMIS-2 |
0.7429 |
0.2571 |
0.4571 |
0.3271 |
0.5486 |
0.3992 |
0.4468 |
KU-DMIS-3 |
0.8286 |
0.2857 |
0.4286 |
0.3452 |
0.5653 |
0.4131 |
0.4619 |
KU-DMIS-4 |
0.7429 |
0.2286 |
0.4571 |
0.3238 |
0.5069 |
0.3575 |
0.4051 |
KU-DMIS-5 |
0.6571 |
0.2857 |
0.5143 |
0.3638 |
0.5050 |
0.3714 |
0.4124 |
LabZhu,FDU |
0.5429 |
0.2000 |
0.2571 |
0.2238 |
0.2347 |
0.3369 |
0.2647 |
LabZhu_FDU |
0.5429 |
- | - | - |
0.1971 |
0.3417 |
0.2249 |
bioasq_experiments1 |
0.5143 |
0.0571 |
0.1143 |
0.0714 |
0.1667 |
0.0694 |
0.0972 |
bioasq_experiments2 |
0.5143 |
- | - | - |
0.1250 |
0.0556 |
0.0750 |
QA1 |
0.4857 |
0.2571 |
0.3714 |
0.2938 |
0.2051 |
0.5127 |
0.2862 |
limsi-reader-UMLS-r1 |
0.5429 |
- | - | - |
- | - | - |
unipi-quokka-QA-1 |
0.5429 |
0.0857 |
0.1714 |
0.1152 |
0.1713 |
0.5873 |
0.2537 |
unipi-quokka-QA-2 |
0.5429 |
0.0857 |
0.1714 |
0.1152 |
0.1713 |
0.5873 |
0.2537 |
unipi-quokka-QA-3 |
0.6857 |
0.0857 |
0.1714 |
0.1152 |
0.1713 |
0.5873 |
0.2537 |
unipi-quokka-QA-4 |
0.7143 |
0.0857 |
0.1714 |
0.1152 |
0.1713 |
0.5873 |
0.2537 |
unipi-quokka-QA-5 |
0.8000 |
0.0857 |
0.1714 |
0.1152 |
0.1713 |
0.5873 |
0.2537 |
BioASQ_Baseline |
0.4857 |
0.0571 |
0.1429 |
0.0867 |
0.2127 |
0.3619 |
0.2573 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
auth-qa-1 |
- | - |
auth-qa-2 |
- | - |
Neural MSE |
0.3765 |
0.3223 |
0.3735 |
0.3147 |
First K sentences |
0.3553 |
0.3038 |
0.3533 |
0.2956 |
auth-qa-3 |
- | - |
MQ-1 |
0.5112 |
0.3445 |
0.5142 |
0.3352 |
MQ-2 |
0.5103 |
0.3362 |
0.5126 |
0.3255 |
MQ-5 |
0.4617 |
0.3351 |
0.4611 |
0.3251 |
MQ-3 |
0.5112 |
0.3445 |
0.5142 |
0.3352 |
auth-qa-4 |
- | - |
MQ-4 |
0.5198 |
0.3504 |
0.5218 |
0.3402 |
google-gold-input-ab |
- | - |
google-pred-input |
- | - |
google-gold-input-nq |
- | - |
google-gold-input |
- | - |
Lab Zhu ,Fdan Univer |
0.4093 |
0.3323 |
0.4103 |
0.3250 |
System B process |
0.1960 |
0.1990 |
0.1930 |
0.1949 |
BJUTNLPGroup |
0.0722 |
0.1027 |
0.0575 |
0.0801 |
UNCC_QA_1 |
- | - |
UNCC_QA3 |
- | - |
UNCC_QA2 |
- | - |
System B+C proc |
0.2066 |
0.1763 |
0.2096 |
0.1772 |
System D+E process |
0.1956 |
0.2048 |
0.1858 |
0.1940 |
System384 B process |
0.2173 |
0.2214 |
0.2098 |
0.2139 |
System384 D+E proces |
0.2053 |
0.2183 |
0.1964 |
0.2096 |
limsi-reader |
- | - |
Neural MSE Attention |
0.3765 |
0.3223 |
0.3735 |
0.3147 |
L2PS - DeepQA |
- | - |
List only |
- | - |
limsi-reader-UMLS-r2 |
- | - |
KU-DMIS-1 |
- | - |
Lab Zhu,Fudan Univer |
0.4093 |
0.3323 |
0.4103 |
0.3250 |
KU-DMIS-2 |
- | - |
KU-DMIS-3 |
- | - |
KU-DMIS-4 |
- | - |
KU-DMIS-5 |
- | - |
LabZhu,FDU |
0.4093 |
0.3323 |
0.4103 |
0.3250 |
LabZhu_FDU |
0.4093 |
0.3323 |
0.4103 |
0.3250 |
bioasq_experiments1 |
0.5056 |
0.3299 |
0.5097 |
0.3200 |
bioasq_experiments2 |
0.5056 |
0.3299 |
0.5097 |
0.3200 |
QA1 |
- | - |
limsi-reader-UMLS-r1 |
- | - |
unipi-quokka-QA-1 |
0.4242 |
0.3104 |
0.4234 |
0.3006 |
unipi-quokka-QA-2 |
0.4242 |
0.3104 |
0.4234 |
0.3006 |
unipi-quokka-QA-3 |
0.4242 |
0.3104 |
0.4234 |
0.3006 |
unipi-quokka-QA-4 |
0.4242 |
0.3104 |
0.4234 |
0.3006 |
unipi-quokka-QA-5 |
0.4242 |
0.3104 |
0.4234 |
0.3006 |
BioASQ_Baseline |
- | - |
+ Task 6b, Test batch 11
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
pa-base |
0.6800 |
0.2188 |
0.5000 |
0.3542 |
0.2750 |
0.2250 |
0.2305 |
pa |
0.6800 |
0.2188 |
0.5000 |
0.3542 |
0.3884 |
0.5629 |
0.4315 |
AUEB-System1 |
0.6800 |
0.0313 |
0.0625 |
0.0469 |
0.0333 |
0.0417 |
0.0367 |
AUEB-System3 |
0.6800 |
- | - | - |
0.0200 |
0.0333 |
0.0250 |
AUEB-System4 |
0.6800 |
0.0313 |
0.0625 |
0.0469 |
0.0500 |
0.0750 |
0.0597 |
AUEB-System5 |
0.6800 |
0.0313 |
0.1563 |
0.0714 |
0.1008 |
0.1308 |
0.1095 |
UoT_baseline |
0.6800 |
0.3125 |
0.5938 |
0.4266 |
0.3461 |
0.2933 |
0.2918 |
UoT_allquestions |
0.6800 |
0.2813 |
0.5938 |
0.4099 |
0.3842 |
0.3200 |
0.3262 |
BJUTNLPGroup |
0.6800 |
0.2813 |
0.4063 |
0.3307 |
0.1800 |
0.5458 |
0.2652 |
MQ-5 |
0.6800 |
- | - | - |
- | - | - |
MQ-1 |
0.6800 |
- | - | - |
- | - | - |
MQ-2 |
0.6800 |
- | - | - |
- | - | - |
MQ-3 |
0.6800 |
- | - | - |
- | - | - |
bio-answerfinder |
0.6800 |
0.0625 |
0.2188 |
0.1406 |
0.1511 |
0.3671 |
0.1673 |
Umass_czi_1 |
0.6800 |
0.3750 |
0.5938 |
0.4688 |
- | - | - |
MQ-4 |
0.6800 |
- | - | - |
- | - | - |
Umass_czi_2 |
0.5600 |
0.2500 |
0.3438 |
0.2891 |
0.4875 |
0.2983 |
0.3448 |
Umass_czi_3 |
0.6800 |
0.2500 |
0.3438 |
0.2891 |
0.4875 |
0.2983 |
0.3448 |
Umass_czi_4 |
0.6400 |
0.2188 |
0.4375 |
0.3005 |
0.4875 |
0.2983 |
0.3448 |
NCU-IISR_1 |
0.7600 |
- | - | - |
- | - | - |
Umass_czi_5 |
0.6400 |
0.2188 |
0.4375 |
0.3005 |
0.1350 |
0.0767 |
0.0967 |
FudanLabZhu1 |
0.6000 |
0.3750 |
0.5938 |
0.4557 |
0.4250 |
0.3067 |
0.3408 |
FudanLabZhu4 |
0.6000 |
0.3125 |
0.5938 |
0.4219 |
0.4250 |
0.3067 |
0.3408 |
auth-qa-1 |
0.6800 |
0.0625 |
0.1563 |
0.1094 |
0.0325 |
0.0583 |
0.0417 |
kmeans |
- |
- | - | - |
- | - | - |
KoreaUniv-DMIS-2 |
0.8400 |
0.3438 |
0.6250 |
0.4583 |
0.4367 |
0.3100 |
0.3333 |
KoreaUniv-DMIS-3 |
0.8800 |
0.2500 |
0.6250 |
0.3979 |
0.3908 |
0.3100 |
0.3152 |
simple truncation |
- |
- | - | - |
- | - | - |
KoreaUniv-DMIS-1 |
0.7600 |
0.1250 |
0.4375 |
0.2615 |
0.4333 |
0.3425 |
0.3597 |
KoreaUniv-DMIS-4 |
0.7600 |
0.3125 |
0.6250 |
0.4344 |
0.4450 |
0.3325 |
0.3516 |
KoreaUniv-DMIS-5 |
0.8800 |
0.3438 |
0.5938 |
0.4438 |
0.4733 |
0.3633 |
0.3718 |
BioASQ_Baseline |
0.4000 |
0.1563 |
0.2813 |
0.2016 |
0.1322 |
0.3483 |
0.1839 |
MQ-6 |
0.6800 |
- | - | - |
- | - | - |
BioASQ Baseline ZS |
0.6800 |
0.0625 |
0.2188 |
0.1406 |
0.1511 |
0.3671 |
0.1673 |
BioASQ_Baseline |
0.6800 |
0.0625 |
0.2188 |
0.1406 |
0.1511 |
0.3671 |
0.1673 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
pa-base |
0.1040 |
0.1061 |
0.1036 |
0.1058 |
pa |
0.1040 |
0.1061 |
0.1036 |
0.1058 |
AUEB-System1 |
0.0944 |
0.0495 |
0.0978 |
0.0496 |
AUEB-System3 |
0.0563 |
0.0278 |
0.0636 |
0.0312 |
AUEB-System4 |
0.0163 |
0.0056 |
0.0246 |
0.0099 |
AUEB-System5 |
0.0595 |
0.0282 |
0.0659 |
0.0317 |
UoT_baseline |
- | - |
UoT_allquestions |
- | - |
BJUTNLPGroup |
0.0794 |
0.0976 |
0.0633 |
0.0731 |
MQ-5 |
0.4616 |
0.3331 |
0.4740 |
0.3295 |
MQ-1 |
0.5322 |
0.3339 |
0.5412 |
0.3276 |
MQ-2 |
0.5671 |
0.3487 |
0.5732 |
0.3417 |
MQ-3 |
0.5789 |
0.3568 |
0.5834 |
0.3481 |
bio-answerfinder |
0.4515 |
0.3289 |
0.4543 |
0.3236 |
Umass_czi_1 |
- | - |
MQ-4 |
0.5401 |
0.3365 |
0.5492 |
0.3301 |
Umass_czi_2 |
- | - |
Umass_czi_3 |
- | - |
Umass_czi_4 |
- | - |
NCU-IISR_1 |
0.1699 |
0.1923 |
0.1735 |
0.1935 |
Umass_czi_5 |
- | - |
FudanLabZhu1 |
- | - |
FudanLabZhu4 |
- | - |
auth-qa-1 |
- | - |
kmeans |
0.5566 |
0.3208 |
0.5609 |
0.3129 |
KoreaUniv-DMIS-2 |
- | - |
KoreaUniv-DMIS-3 |
- | - |
simple truncation |
0.3841 |
0.3385 |
0.3896 |
0.3339 |
KoreaUniv-DMIS-1 |
0.2326 |
0.1576 |
0.2509 |
0.1651 |
KoreaUniv-DMIS-4 |
0.1957 |
0.1887 |
0.2021 |
0.1898 |
KoreaUniv-DMIS-5 |
0.1960 |
0.1895 |
0.2019 |
0.1902 |
BioASQ_Baseline |
- | - |
MQ-6 |
0.4048 |
0.0000 |
0.5985 |
0.3103 |
BioASQ Baseline ZS |
0.4048 |
0.0000 |
0.4515 |
0.3289 |
BioASQ_Baseline |
0.4048 |
0.0000 |
0.4515 |
0.3289 |
+ Task 6b, Test batch 12
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
BJUTNLPGroup |
0.7500 |
0.0800 |
0.3200 |
0.1747 |
0.1429 |
0.4619 |
0.2027 |
AUEB-System1 |
0.7500 |
- | - | - |
0.0179 |
0.0179 |
0.0179 |
AUEB-System2 |
0.7500 |
- | - | - |
0.0143 |
0.0238 |
0.0179 |
AUEB-System3 |
0.7500 |
- | - | - |
- | - | - |
AUEB-System4 |
0.7500 |
- | - | - |
0.0429 |
0.0524 |
0.0429 |
AUEB-System5 |
0.7500 |
- | - | - |
0.0667 |
0.0821 |
0.0672 |
auth-qa-1 |
0.7778 |
- | - | - |
0.0845 |
0.1238 |
0.0944 |
bio-answerfinder |
0.7778 |
0.2000 |
0.2400 |
0.2080 |
0.3693 |
0.4714 |
0.3803 |
UoT_baseline |
0.7500 |
0.1600 |
0.4400 |
0.2580 |
0.5361 |
0.4476 |
0.4306 |
Best yesno |
0.7500 |
0.1600 |
0.4400 |
0.2580 |
0.5361 |
0.4476 |
0.4306 |
UoT_multitask_learn |
0.8333 |
0.2000 |
0.4000 |
0.2800 |
0.4643 |
0.4214 |
0.4108 |
UoT_allquestions |
0.7778 |
0.1600 |
0.4800 |
0.2540 |
0.4107 |
0.3738 |
0.3712 |
MQ-1 |
- |
- | - | - |
- | - | - |
Best factoid |
0.7500 |
0.1200 |
0.4400 |
0.2413 |
0.4827 |
0.4071 |
0.3950 |
Umass_czi_2 |
0.7778 |
0.0800 |
0.2400 |
0.1333 |
- | - | - |
Umass_czi_1 |
0.7222 |
0.1200 |
0.2000 |
0.1480 |
- | - | - |
NCU-IISR_1 |
0.7778 |
0.2800 |
0.4400 |
0.3293 |
0.3214 |
0.2381 |
0.2667 |
MQ-2 |
0.7500 |
- | - | - |
- | - | - |
MQ-3 |
0.7500 |
- | - | - |
- | - | - |
MQ-4 |
0.7500 |
- | - | - |
- | - | - |
simple truncation |
- |
- | - | - |
- | - | - |
MQ-5 |
0.7500 |
- | - | - |
- | - | - |
Multitask SBERT Cls |
0.7500 |
- | - | - |
- | - | - |
Multitask SBERT reg |
0.7500 |
- | - | - |
- | - | - |
sbert cls |
0.7500 |
- | - | - |
- | - | - |
sbert reg |
0.7500 |
- | - | - |
- | - | - |
pa-base |
0.2500 |
0.1200 |
0.4000 |
0.2300 |
0.1012 |
0.0619 |
0.0702 |
FudanLabZhu1 |
0.6944 |
0.2800 |
0.4000 |
0.3200 |
0.3631 |
0.2976 |
0.3112 |
FudanLabZhu3 |
0.6944 |
0.2000 |
0.3200 |
0.2413 |
0.5417 |
0.5024 |
0.4678 |
kmeans |
- |
- | - | - |
- | - | - |
FudanLabZhu4 |
0.6944 |
0.2400 |
0.3600 |
0.2900 |
0.5417 |
0.5024 |
0.4678 |
KoreaUniv-DMIS-1 |
0.9444 |
0.1200 |
0.3200 |
0.1967 |
0.5643 |
0.4643 |
0.4735 |
KoreaUniv-DMIS-2 |
0.9167 |
0.1600 |
0.3600 |
0.2367 |
0.5060 |
0.3667 |
0.4029 |
KoreaUniv-DMIS-5 |
0.8889 |
0.1600 |
0.3600 |
0.2193 |
0.4821 |
0.3667 |
0.3985 |
KoreaUniv-DMIS-3 |
0.9444 |
0.1200 |
0.3200 |
0.2000 |
0.5500 |
0.4214 |
0.4544 |
KoreaUniv-DMIS-4 |
0.9167 |
0.2800 |
0.4400 |
0.3533 |
0.4881 |
0.3381 |
0.3798 |
pa |
0.2500 |
0.1200 |
0.4400 |
0.2360 |
0.3418 |
0.4881 |
0.3441 |
BioASQ_Baseline |
0.3611 |
0.0800 |
0.1200 |
0.1000 |
0.2196 |
0.5929 |
0.2858 |
MQ-6 |
0.7500 |
- | - | - |
- | - | - |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
BJUTNLPGroup |
0.0325 |
0.0555 |
0.0223 |
0.0389 |
AUEB-System1 |
0.0551 |
0.0331 |
0.0590 |
0.0338 |
AUEB-System2 |
0.0491 |
0.0289 |
0.0550 |
0.0309 |
AUEB-System3 |
0.0551 |
0.0331 |
0.0590 |
0.0338 |
AUEB-System4 |
0.2125 |
0.1151 |
0.2113 |
0.1128 |
AUEB-System5 |
0.2125 |
0.1151 |
0.2113 |
0.1128 |
auth-qa-1 |
- | - |
bio-answerfinder |
0.4150 |
0.2997 |
0.4231 |
0.2981 |
UoT_baseline |
- | - |
Best yesno |
- | - |
UoT_multitask_learn |
- | - |
UoT_allquestions |
- | - |
MQ-1 |
0.5205 |
0.3040 |
0.5243 |
0.2939 |
Best factoid |
- | - |
Umass_czi_2 |
- | - |
Umass_czi_1 |
- | - |
NCU-IISR_1 |
0.1560 |
0.1841 |
0.1547 |
0.1783 |
MQ-2 |
0.5244 |
0.3254 |
0.5339 |
0.3206 |
MQ-3 |
0.5232 |
0.3206 |
0.5314 |
0.3153 |
MQ-4 |
0.4782 |
0.3012 |
0.4885 |
0.2975 |
simple truncation |
0.5201 |
0.3011 |
0.5228 |
0.2906 |
MQ-5 |
0.3794 |
0.2683 |
0.3885 |
0.2651 |
Multitask SBERT Cls |
0.4701 |
0.2992 |
0.4813 |
0.2947 |
Multitask SBERT reg |
0.4649 |
0.2938 |
0.4747 |
0.2898 |
sbert cls |
0.4806 |
0.2993 |
0.4915 |
0.2954 |
sbert reg |
0.4593 |
0.2904 |
0.4784 |
0.2918 |
pa-base |
0.0592 |
0.0653 |
0.0590 |
0.0642 |
FudanLabZhu1 |
- | - |
FudanLabZhu3 |
- | - |
kmeans |
0.5138 |
0.2891 |
0.5281 |
0.2845 |
FudanLabZhu4 |
- | - |
KoreaUniv-DMIS-1 |
0.2387 |
0.2287 |
0.2486 |
0.2311 |
KoreaUniv-DMIS-2 |
0.2546 |
0.2442 |
0.2567 |
0.2413 |
KoreaUniv-DMIS-5 |
0.2402 |
0.2318 |
0.2486 |
0.2335 |
KoreaUniv-DMIS-3 |
0.2662 |
0.2553 |
0.2687 |
0.2527 |
KoreaUniv-DMIS-4 |
0.2363 |
0.2272 |
0.2494 |
0.2331 |
pa |
0.0592 |
0.0653 |
0.0590 |
0.0642 |
BioASQ_Baseline |
- | - |
MQ-6 |
0.4286 |
0.0000 |
0.5544 |
0.2929 |
+ Task 6b, Test batch 13
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
bio-answerfinder |
0.8710 |
0.3214 |
0.4286 |
0.3494 |
0.3884 |
0.4972 |
0.3983 |
auth-qa-1 |
0.7742 |
0.2143 |
0.2857 |
0.2411 |
0.2500 |
0.3917 |
0.2834 |
auth-qa-2 |
0.7742 |
0.2143 |
0.2857 |
0.2440 |
0.1667 |
0.4722 |
0.2337 |
auth-qa-3 |
0.7742 |
0.2143 |
0.2857 |
0.2440 |
0.2500 |
0.3917 |
0.2834 |
auth-qa-4 |
0.7742 |
0.2143 |
0.2857 |
0.2411 |
0.1667 |
0.4722 |
0.2337 |
MQ-1 |
0.5806 |
- | - | - |
- | - | - |
MQ-2 |
0.5806 |
- | - | - |
- | - | - |
MQ-3 |
0.5806 |
- | - | - |
- | - | - |
MQ-4 |
0.5806 |
- | - | - |
- | - | - |
MQ-5 |
0.5806 |
- | - | - |
- | - | - |
Umass_czi_1 |
0.8065 |
0.2500 |
0.3571 |
0.2869 |
0.6806 |
0.4111 |
0.4474 |
Umass_czi_2 |
0.8387 |
0.2500 |
0.3571 |
0.2869 |
0.6806 |
0.4111 |
0.4474 |
Umass_czi_4 |
0.9032 |
0.3214 |
0.4643 |
0.3810 |
0.6111 |
0.4028 |
0.4314 |
Umass_czi_5 |
0.9032 |
0.2500 |
0.4286 |
0.3030 |
0.7361 |
0.4500 |
0.5020 |
UoT_baseline |
0.5806 |
0.3214 |
0.3929 |
0.3512 |
0.4861 |
0.3833 |
0.4024 |
UoT_allquestions |
0.5806 |
0.3214 |
0.3929 |
0.3423 |
0.5972 |
0.3889 |
0.4151 |
UoT_multitask_learn |
0.5161 |
0.3214 |
0.4286 |
0.3643 |
0.5139 |
0.3333 |
0.3530 |
Best factoid |
0.5806 |
0.2857 |
0.3929 |
0.3333 |
0.5208 |
0.3833 |
0.3917 |
Best yesno |
0.5161 |
0.3214 |
0.4286 |
0.3643 |
0.5139 |
0.3333 |
0.3530 |
BJUTNLPGroup |
0.5806 |
0.2500 |
0.3571 |
0.3036 |
0.1583 |
0.4222 |
0.2170 |
Multitask SBERT Cls |
0.5806 |
- | - | - |
- | - | - |
Multitask SBERT reg |
0.5806 |
- | - | - |
- | - | - |
sbert cls |
0.5806 |
- | - | - |
- | - | - |
sbert 1 epoch cls |
0.5806 |
- | - | - |
- | - | - |
GNN |
0.5806 |
- | - | - |
- | - | - |
factoid qa model |
0.4194 |
0.3214 |
0.4643 |
0.3750 |
- | - | - |
simple truncation |
0.5806 |
- | - | - |
- | - | - |
kmeans |
0.5806 |
- | - | - |
- | - | - |
similarity measures |
0.5806 |
- | - | - |
- | - | - |
abstractive |
0.5806 |
- | - | - |
- | - | - |
extractive |
0.5806 |
- | - | - |
- | - | - |
pa-base |
0.9032 |
0.2500 |
0.4643 |
0.3137 |
0.5278 |
0.4444 |
0.4377 |
pa |
0.9032 |
0.2500 |
0.4643 |
0.3137 |
0.5278 |
0.4444 |
0.4377 |
KoreaUniv-DMIS-1 |
0.9032 |
0.3214 |
0.4286 |
0.3601 |
0.6583 |
0.4111 |
0.4312 |
KoreaUniv-DMIS-4 |
0.8387 |
0.2857 |
0.4286 |
0.3357 |
0.6167 |
0.4111 |
0.4282 |
KoreaUniv-DMIS-2 |
0.8710 |
0.3214 |
0.4286 |
0.3446 |
0.6028 |
0.4111 |
0.4259 |
KoreaUniv-DMIS-3 |
0.8387 |
0.2500 |
0.4643 |
0.3357 |
0.6111 |
0.4111 |
0.4222 |
KoreaUniv-DMIS-5 |
0.9032 |
0.3214 |
0.4643 |
0.3565 |
0.6167 |
0.4111 |
0.4282 |
sbert reg |
0.5806 |
- | - | - |
- | - | - |
FudanLabZhu1 |
0.7419 |
0.2500 |
0.3929 |
0.2976 |
0.4819 |
0.2806 |
0.3112 |
FudanLabZhu2 |
0.7419 |
0.3214 |
0.5357 |
0.3970 |
0.5694 |
0.3361 |
0.3849 |
FudanLabZhu3 |
0.7419 |
0.3214 |
0.4643 |
0.3655 |
0.5583 |
0.3361 |
0.3708 |
FudanLabZhu4 |
0.7419 |
0.2857 |
0.5714 |
0.3821 |
0.5583 |
0.3361 |
0.3708 |
FudanLabZhu5 |
0.7419 |
0.3214 |
0.4286 |
0.3690 |
0.5583 |
0.3361 |
0.3708 |
BioASQ_Baseline |
0.5161 |
0.0714 |
0.2143 |
0.1220 |
0.2052 |
0.4611 |
0.2456 |
MQ-6 |
0.5806 |
- | - | - |
- | - | - |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
bio-answerfinder |
0.4557 |
0.3471 |
0.4642 |
0.3450 |
auth-qa-1 |
- | - |
auth-qa-2 |
- | - |
auth-qa-3 |
- | - |
auth-qa-4 |
- | - |
MQ-1 |
0.5222 |
0.3632 |
0.5336 |
0.3597 |
MQ-2 |
0.5481 |
0.3719 |
0.5580 |
0.3673 |
MQ-3 |
0.5394 |
0.3673 |
0.5491 |
0.3633 |
MQ-4 |
0.5202 |
0.3597 |
0.5336 |
0.3574 |
MQ-5 |
0.3997 |
0.3055 |
0.4112 |
0.3060 |
Umass_czi_1 |
- | - |
Umass_czi_2 |
- | - |
Umass_czi_4 |
- | - |
Umass_czi_5 |
- | - |
UoT_baseline |
- | - |
UoT_allquestions |
- | - |
UoT_multitask_learn |
- | - |
Best factoid |
- | - |
Best yesno |
- | - |
BJUTNLPGroup |
0.0293 |
0.0501 |
0.0196 |
0.0337 |
Multitask SBERT Cls |
0.4812 |
0.3292 |
0.5007 |
0.3305 |
Multitask SBERT reg |
0.4669 |
0.3166 |
0.4834 |
0.3178 |
sbert cls |
0.4769 |
0.3177 |
0.4940 |
0.3189 |
sbert 1 epoch cls |
0.4820 |
0.3225 |
0.4977 |
0.3232 |
GNN |
0.1538 |
0.1453 |
0.1600 |
0.1439 |
factoid qa model |
- | - |
simple truncation |
0.4154 |
0.3229 |
0.4256 |
0.3221 |
kmeans |
0.4154 |
0.3229 |
0.4256 |
0.3221 |
similarity measures |
0.3961 |
0.2941 |
0.4093 |
0.2959 |
abstractive |
0.1959 |
0.2186 |
0.1943 |
0.2169 |
extractive |
0.4519 |
0.3197 |
0.4629 |
0.3155 |
pa-base |
0.3202 |
0.2896 |
0.3261 |
0.2921 |
pa |
0.5088 |
0.3229 |
0.5179 |
0.3182 |
KoreaUniv-DMIS-1 |
0.3213 |
0.2832 |
0.3193 |
0.2787 |
KoreaUniv-DMIS-4 |
0.2704 |
0.2539 |
0.2701 |
0.2521 |
KoreaUniv-DMIS-2 |
0.2992 |
0.2809 |
0.2983 |
0.2795 |
KoreaUniv-DMIS-3 |
0.2854 |
0.2660 |
0.2852 |
0.2634 |
KoreaUniv-DMIS-5 |
0.2997 |
0.2899 |
0.3043 |
0.2909 |
sbert reg |
0.4508 |
0.3110 |
0.4647 |
0.3121 |
FudanLabZhu1 |
- | - |
FudanLabZhu2 |
- | - |
FudanLabZhu3 |
- | - |
FudanLabZhu4 |
- | - |
FudanLabZhu5 |
- | - |
BioASQ_Baseline |
- | - |
MQ-6 |
0.3673 |
0.0000 |
0.5269 |
0.3056 |
+ Task 6b, Test batch 14
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
MQ-2 |
0.5385 |
- | - | - |
- | - | - |
MQ-3 |
0.5385 |
- | - | - |
- | - | - |
MQ-4 |
0.5385 |
- | - | - |
- | - | - |
bio-answerfinder |
0.6538 |
0.4706 |
0.5882 |
0.5245 |
0.2791 |
0.3636 |
0.2646 |
auth-qa-1 |
0.6154 |
0.2941 |
0.5294 |
0.3794 |
0.3059 |
0.4229 |
0.3406 |
auth-qa-2 |
0.6154 |
0.2941 |
0.5000 |
0.3647 |
0.1882 |
0.4871 |
0.2561 |
auth-qa-3 |
0.6154 |
0.2941 |
0.5000 |
0.3647 |
0.3137 |
0.4283 |
0.3459 |
auth-qa-4 |
0.6154 |
0.2941 |
0.5294 |
0.3794 |
0.1882 |
0.4871 |
0.2561 |
auth-qa-5 |
0.6154 |
0.2941 |
0.5294 |
0.3794 |
0.1882 |
0.4871 |
0.2561 |
MQ-1 |
0.5385 |
- | - | - |
- | - | - |
MQ-5 |
0.5385 |
- | - | - |
- | - | - |
BJUTNLPGroup |
0.5385 |
0.3529 |
0.5000 |
0.4167 |
0.1824 |
0.5597 |
0.2620 |
UoT_baseline |
0.5000 |
0.4118 |
0.7059 |
0.5270 |
0.3296 |
0.3810 |
0.3161 |
UoT_allquestions |
0.5000 |
0.4412 |
0.7059 |
0.5564 |
0.4045 |
0.4623 |
0.3886 |
Best factoid |
0.5000 |
0.4706 |
0.7059 |
0.5564 |
0.4582 |
0.4153 |
0.4005 |
UoT_multitask_learn |
0.4615 |
0.4706 |
0.6765 |
0.5637 |
0.3843 |
0.3226 |
0.2991 |
Best yesno |
0.5000 |
0.4412 |
0.7059 |
0.5564 |
0.4045 |
0.4623 |
0.3886 |
GNN |
0.5385 |
0.0294 |
0.2353 |
0.1039 |
- | - | - |
Umass_czi_1 |
0.8077 |
0.3529 |
0.5294 |
0.4186 |
0.5164 |
0.3888 |
0.3774 |
Umass_czi_2 |
0.6923 |
0.3235 |
0.5000 |
0.3946 |
0.5164 |
0.3888 |
0.3774 |
Umass_czi_3 |
0.7692 |
0.3235 |
0.5000 |
0.3931 |
0.5753 |
0.4182 |
0.4146 |
Umass_czi_4 |
0.6538 |
0.2941 |
0.5588 |
0.3946 |
0.5753 |
0.4182 |
0.4146 |
Umass_czi_5 |
0.6538 |
0.5000 |
0.7059 |
0.5637 |
0.5753 |
0.4182 |
0.4146 |
factoid qa model |
0.7308 |
0.4412 |
0.6471 |
0.5206 |
0.4039 |
0.1696 |
0.2117 |
Parameters retrained |
0.7308 |
0.4412 |
0.6765 |
0.5216 |
0.4679 |
0.3443 |
0.3341 |
Features Fusion |
0.7308 |
0.5000 |
0.6765 |
0.5745 |
0.5428 |
0.3541 |
0.3625 |
Multitask SBERT Cls |
0.5385 |
- | - | - |
- | - | - |
Multitask SBERT reg |
0.5385 |
- | - | - |
- | - | - |
sbert cls |
0.5385 |
- | - | - |
- | - | - |
sbert reg |
0.5385 |
- | - | - |
- | - | - |
sbert 1 epoch cls |
0.5385 |
- | - | - |
- | - | - |
NCU-IISR_1 |
0.7308 |
0.5000 |
0.6765 |
0.5735 |
0.5539 |
0.3786 |
0.3905 |
dice-a-1.0 |
0.7308 |
0.3824 |
0.6471 |
0.4926 |
0.3627 |
0.1965 |
0.2188 |
FudanLabZhu1 |
0.5769 |
0.4118 |
0.5588 |
0.4804 |
0.4055 |
0.2970 |
0.3004 |
FudanLabZhu2 |
0.5769 |
0.5294 |
0.6765 |
0.5980 |
0.5784 |
0.3541 |
0.3902 |
FudanLabZhu4 |
0.5769 |
0.5000 |
0.6765 |
0.5686 |
0.5375 |
0.5089 |
0.4571 |
FudanLabZhu3 |
0.5769 |
0.4118 |
0.5882 |
0.4755 |
0.5375 |
0.5089 |
0.4571 |
GIAO |
0.6538 |
0.5000 |
0.6765 |
0.5784 |
0.6520 |
0.3585 |
0.4101 |
KoreaUniv-DMIS-1 |
0.7692 |
0.5294 |
0.7059 |
0.6078 |
0.3577 |
0.5539 |
0.4037 |
KoreaUniv-DMIS-2 |
0.7692 |
0.5294 |
0.6765 |
0.5882 |
0.2760 |
0.4926 |
0.3122 |
dice-b-1.0 |
0.7308 |
0.3824 |
0.6471 |
0.4926 |
0.3775 |
0.2259 |
0.2384 |
simple truncation |
0.5385 |
- | - | - |
- | - | - |
kmeans |
0.5385 |
- | - | - |
- | - | - |
similarity measures |
0.5385 |
- | - | - |
- | - | - |
KoreaUniv-DMIS-3 |
0.7692 |
0.5000 |
0.6765 |
0.5706 |
0.3049 |
0.4461 |
0.3346 |
KoreaUniv-DMIS-4 |
0.8077 |
0.5000 |
0.6471 |
0.5613 |
0.3002 |
0.4926 |
0.3318 |
extractive |
0.5385 |
- | - | - |
- | - | - |
abstractive |
0.5385 |
- | - | - |
- | - | - |
FudanLabZhu5 |
0.5769 |
0.5588 |
0.7353 |
0.6284 |
0.5375 |
0.5089 |
0.4571 |
KoreaUniv-DMIS-5 |
0.8462 |
0.4706 |
0.7059 |
0.5686 |
0.3630 |
0.4329 |
0.3355 |
pa |
0.7308 |
0.4706 |
0.5588 |
0.5098 |
0.3571 |
0.3661 |
0.3030 |
pa-base |
0.7308 |
0.4706 |
0.5588 |
0.5098 |
0.3571 |
0.3661 |
0.3030 |
BioASQ_Baseline |
0.5385 |
0.0588 |
0.2059 |
0.1078 |
0.1554 |
0.4519 |
0.2122 |
BioASQ_Baseline |
0.5385 |
0.0588 |
0.2059 |
0.1078 |
0.1554 |
0.4519 |
0.2122 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
MQ-2 |
0.5054 |
0.2949 |
0.5122 |
0.2878 |
MQ-3 |
0.5162 |
0.2964 |
0.5220 |
0.2891 |
MQ-4 |
0.4915 |
0.2886 |
0.5015 |
0.2832 |
bio-answerfinder |
0.4025 |
0.2993 |
0.4001 |
0.2938 |
auth-qa-1 |
- | - |
auth-qa-2 |
- | - |
auth-qa-3 |
- | - |
auth-qa-4 |
- | - |
auth-qa-5 |
- | - |
MQ-1 |
0.4971 |
0.3088 |
0.5074 |
0.3017 |
MQ-5 |
0.3896 |
0.2674 |
0.4005 |
0.2638 |
BJUTNLPGroup |
0.0288 |
0.0488 |
0.0184 |
0.0316 |
UoT_baseline |
- | - |
UoT_allquestions |
- | - |
Best factoid |
- | - |
UoT_multitask_learn |
- | - |
Best yesno |
- | - |
GNN |
0.1543 |
0.1385 |
0.1538 |
0.1330 |
Umass_czi_1 |
- | - |
Umass_czi_2 |
- | - |
Umass_czi_3 |
- | - |
Umass_czi_4 |
- | - |
Umass_czi_5 |
- | - |
factoid qa model |
- | - |
Parameters retrained |
- | - |
Features Fusion |
- | - |
Multitask SBERT Cls |
0.4195 |
0.2492 |
0.4282 |
0.2453 |
Multitask SBERT reg |
0.4301 |
0.2582 |
0.4414 |
0.2538 |
sbert cls |
0.4294 |
0.2575 |
0.4382 |
0.2534 |
sbert reg |
0.4294 |
0.2575 |
0.4382 |
0.2534 |
sbert 1 epoch cls |
0.3996 |
0.2502 |
0.4147 |
0.2482 |
NCU-IISR_1 |
0.1616 |
0.1845 |
0.1587 |
0.1796 |
dice-a-1.0 |
- | - |
FudanLabZhu1 |
- | - |
FudanLabZhu2 |
- | - |
FudanLabZhu4 |
- | - |
FudanLabZhu3 |
- | - |
GIAO |
- | - |
KoreaUniv-DMIS-1 |
0.2390 |
0.2136 |
0.2436 |
0.2141 |
KoreaUniv-DMIS-2 |
0.2423 |
0.2274 |
0.2456 |
0.2287 |
dice-b-1.0 |
- | - |
simple truncation |
0.4793 |
0.2828 |
0.4929 |
0.2782 |
kmeans |
0.4793 |
0.2828 |
0.4929 |
0.2782 |
similarity measures |
0.3862 |
0.2446 |
0.3971 |
0.2394 |
KoreaUniv-DMIS-3 |
0.2407 |
0.2246 |
0.2424 |
0.2251 |
KoreaUniv-DMIS-4 |
0.2268 |
0.2061 |
0.2291 |
0.2066 |
extractive |
0.4668 |
0.2870 |
0.4729 |
0.2796 |
abstractive |
0.1938 |
0.2080 |
0.1948 |
0.2050 |
FudanLabZhu5 |
- | - |
KoreaUniv-DMIS-5 |
0.2302 |
0.2126 |
0.2384 |
0.2181 |
pa |
0.3135 |
0.2904 |
0.3138 |
0.2874 |
pa-base |
0.5291 |
0.2923 |
0.5321 |
0.2856 |
BioASQ_Baseline |
- | - |
BioASQ_Baseline |
0.5125 |
0.6250 |
0.0000 |
0.0000 |
+ Task 6b, Test batch 15
Exact Answers
|
Yes/No |
Factoid |
List |
System Name |
Accuracy |
Strict Acc. |
Lenient Acc. |
MRR |
Mean precision |
Recall |
F-Measure |
MQ-2 |
0.5588 |
- | - | - |
- | - | - |
MQ-3 |
0.5588 |
- | - | - |
- | - | - |
MQ-4 |
0.5588 |
- | - | - |
- | - | - |
GNN |
0.5588 |
- | - | - |
- | - | - |
zmodel2 |
0.5588 |
- | - | - |
- | - | - |
NCU-IISR_1 |
0.7353 |
0.4688 |
0.7188 |
0.5859 |
0.4514 |
0.2659 |
0.3140 |
NCU-IISR_2 |
0.5588 |
- | - | - |
- | - | - |
auth-qa-1 |
0.6765 |
0.2500 |
0.3750 |
0.2995 |
0.1750 |
0.4821 |
0.2386 |
auth-qa-5 |
0.6471 |
0.2500 |
0.3750 |
0.2995 |
0.1750 |
0.4821 |
0.2386 |
UoT_baseline |
0.6176 |
0.5000 |
0.6875 |
0.5844 |
0.2242 |
0.1577 |
0.1732 |
UoT_multitask_learn |
0.5000 |
0.4063 |
0.7188 |
0.5365 |
0.5938 |
0.3700 |
0.4296 |
UoT_allquestions |
0.5588 |
0.4063 |
0.6875 |
0.5063 |
0.3854 |
0.2798 |
0.3082 |
Best factoid |
0.5588 |
0.4063 |
0.6563 |
0.5026 |
0.5174 |
0.3631 |
0.4002 |
Best yesno |
0.5588 |
0.4063 |
0.6875 |
0.5063 |
0.3854 |
0.2798 |
0.3082 |
bio-answerfinder |
0.7353 |
0.5000 |
0.5313 |
0.5156 |
0.4745 |
0.4325 |
0.4163 |
MQ-1 |
0.5588 |
- | - | - |
- | - | - |
MQ-5 |
0.5588 |
- | - | - |
- | - | - |
BJUTNLP2 |
0.5588 |
0.4688 |
0.5625 |
0.5026 |
0.1250 |
0.3790 |
0.1731 |
Umass_czi_1 |
0.6471 |
0.4688 |
0.7188 |
0.5677 |
0.5139 |
0.2808 |
0.3353 |
Umass_czi_2 |
0.6176 |
0.5000 |
0.6250 |
0.5417 |
0.5139 |
0.2808 |
0.3353 |
Umass_czi_3 |
0.7941 |
0.5625 |
0.7188 |
0.6354 |
0.1528 |
0.1230 |
0.1310 |
Umass_czi_4 |
0.7353 |
0.5313 |
0.6563 |
0.5833 |
0.3750 |
0.1756 |
0.2166 |
Umass_czi_5 |
0.5882 |
0.4688 |
0.7188 |
0.5604 |
0.5972 |
0.3224 |
0.3909 |
factoid qa model |
0.7647 |
0.4688 |
0.6563 |
0.5401 |
0.3333 |
0.0923 |
0.1387 |
Parameters retrained |
0.7647 |
0.4688 |
0.7813 |
0.5938 |
0.5139 |
0.2946 |
0.3492 |
Features Fusion |
0.7647 |
0.5313 |
0.7500 |
0.6115 |
0.5035 |
0.2808 |
0.3298 |
BioFusion |
0.7647 |
0.5000 |
0.7188 |
0.5818 |
0.3854 |
0.2589 |
0.2907 |
BioLabel |
0.7647 |
0.4688 |
0.7188 |
0.5573 |
0.4271 |
0.2798 |
0.3185 |
dice-a-1.0 |
0.6765 |
0.4375 |
0.6250 |
0.5156 |
0.3750 |
0.1696 |
0.2159 |
dice-b-1.0 |
0.7941 |
0.5313 |
0.6875 |
0.5885 |
0.4028 |
0.1845 |
0.2313 |
DAIICT_lex_UMLSgraph |
0.5588 |
- | - | - |
- | - | - |
DAIICT_QSM_UMLSgraph |
0.5588 |
- | - | - |
- | - | - |
DAIICT_QSM |
0.5588 |
- | - | - |
- | - | - |
DAIICT_lex |
0.5588 |
- | - | - |
- | - | - |
system of teamdaiict |
0.5588 |
- | - | - |
- | - | - |
FudanLabZhu1 |
0.5000 |
0.4375 |
0.6250 |
0.5036 |
0.4694 |
0.3304 |
0.3641 |
FudanLabZhu2 |
0.6176 |
0.5000 |
0.7188 |
0.5818 |
0.6458 |
0.4028 |
0.4703 |
FudanLabZhu4 |
0.6176 |
0.1250 |
0.2813 |
0.1901 |
0.6458 |
0.4028 |
0.4703 |
FudanLabZhu3 |
0.6176 |
0.4688 |
0.6250 |
0.5313 |
0.5125 |
0.3869 |
0.4092 |
KoreaUniv-DMIS-1 |
0.8235 |
0.4688 |
0.6250 |
0.5208 |
0.5799 |
0.4812 |
0.5050 |
Multitask SBERT Cls |
0.5588 |
- | - | - |
- | - | - |
KoreaUniv-DMIS-2 |
0.8235 |
0.5000 |
0.7188 |
0.5833 |
0.5799 |
0.4812 |
0.5050 |
Multitask SBERT reg |
0.5588 |
- | - | - |
- | - | - |
KoreaUniv-DMIS-3 |
0.8235 |
0.4688 |
0.7188 |
0.5661 |
0.5694 |
0.5437 |
0.5222 |
KoreaUniv-DMIS-4 |
0.7941 |
0.5313 |
0.7188 |
0.6120 |
0.5724 |
0.5437 |
0.5247 |
KoreaUniv-DMIS-5 |
0.8529 |
0.5000 |
0.6875 |
0.5677 |
0.5465 |
0.5645 |
0.5243 |
sbert cls |
0.5588 |
- | - | - |
- | - | - |
sbert 1 epoch cls |
0.5588 |
- | - | - |
- | - | - |
sbert reg |
0.5588 |
- | - | - |
- | - | - |
FudanLabZhu5 |
0.6176 |
0.5000 |
0.6563 |
0.5677 |
0.5317 |
0.3879 |
0.4239 |
dice-c-1.0 |
0.8529 |
0.3125 |
0.5625 |
0.4151 |
0.1250 |
0.0536 |
0.0741 |
dice-d-1.0 |
0.8529 |
0.3125 |
0.5625 |
0.4151 |
0.1250 |
0.0536 |
0.0741 |
NCU-IISR_3 |
0.5588 |
- | - | - |
- | - | - |
dice-e-1.0 |
0.8529 |
0.3125 |
0.5625 |
0.4151 |
0.1250 |
0.0536 |
0.0741 |
pa |
0.8235 |
0.4375 |
0.6250 |
0.5260 |
0.3284 |
0.2679 |
0.2761 |
GIAO |
0.5588 |
0.5000 |
0.6563 |
0.5677 |
- | - | - |
BioASQ_Baseline |
0.6176 |
0.1563 |
0.3438 |
0.2266 |
0.2573 |
0.3641 |
0.2581 |
Ideal Answers
|
Automatic scores |
System Name |
Rouge-2 |
Rouge-2 - F1 |
Rouge-SU4 |
Rouge-SU4 - F1 |
MQ-2 |
0.5105 |
0.3246 |
0.5161 |
0.3171 |
MQ-3 |
0.5188 |
0.3328 |
0.5214 |
0.3241 |
MQ-4 |
0.5155 |
0.3311 |
0.5188 |
0.3233 |
GNN |
0.2192 |
0.1987 |
0.2151 |
0.1889 |
zmodel2 |
0.1882 |
0.1675 |
0.1855 |
0.1573 |
NCU-IISR_1 |
0.1634 |
0.1793 |
0.1552 |
0.1692 |
NCU-IISR_2 |
0.3027 |
0.2842 |
0.3019 |
0.2760 |
auth-qa-1 |
- | - |
auth-qa-5 |
- | - |
UoT_baseline |
- | - |
UoT_multitask_learn |
- | - |
UoT_allquestions |
- | - |
Best factoid |
- | - |
Best yesno |
- | - |
bio-answerfinder |
0.4057 |
0.2971 |
0.4021 |
0.2892 |
MQ-1 |
0.5050 |
0.3154 |
0.5129 |
0.3074 |
MQ-5 |
0.4069 |
0.3094 |
0.4151 |
0.3051 |
BJUTNLP2 |
0.0373 |
0.0608 |
0.0244 |
0.0413 |
Umass_czi_1 |
- | - |
Umass_czi_2 |
- | - |
Umass_czi_3 |
- | - |
Umass_czi_4 |
- | - |
Umass_czi_5 |
- | - |
factoid qa model |
- | - |
Parameters retrained |
- | - |
Features Fusion |
- | - |
BioFusion |
- | - |
BioLabel |
- | - |
dice-a-1.0 |
- | - |
dice-b-1.0 |
- | - |
DAIICT_lex_UMLSgraph |
0.6250 |
0.3189 |
0.6234 |
0.3059 |
DAIICT_QSM_UMLSgraph |
0.6250 |
0.3189 |
0.6234 |
0.3059 |
DAIICT_QSM |
0.6428 |