commit file to repo
This commit is contained in:
commit
cc6196c5eb
1
.gitattributes
vendored
Normal file
1
.gitattributes
vendored
Normal file
@ -0,0 +1 @@
|
|||||||
|
*.json filter=lfs diff=lfs merge=lfs -text
|
||||||
0
.gitignore
vendored
Normal file
0
.gitignore
vendored
Normal file
1347
configs/20260204_103117.py
Normal file
1347
configs/20260204_103117.py
Normal file
File diff suppressed because it is too large
Load Diff
@ -0,0 +1,7 @@
|
|||||||
|
[RISE-CORE Msg(16240:140016230472704:libvgpu.c:901)]: Initializing.....
|
||||||
|
[RISE-CORE ERROR (pid:16240 thread=140016230472704 libvgpu.c:958)]: cuInit failed:100
|
||||||
|
02/04 11:12:02 - OpenCompass - INFO - Task [public/qwen3-0-6b@main/GaokaoBench_2010-2013_English_MCQs]: {'score': 40.0}
|
||||||
|
02/04 11:12:02 - OpenCompass - INFO - time elapsed: 2.30s
|
||||||
|
/opt/conda/lib/python3.8/site-packages/fuzzywuzzy/fuzz.py:11: UserWarning: Using slow pure-python SequenceMatcher. Install python-Levenshtein to remove this warning
|
||||||
|
warnings.warn('Using slow pure-python SequenceMatcher. Install python-Levenshtein to remove this warning')
|
||||||
|
[RISE-CORE Msg(16240:140016230472704:multiprocess_memory_limit.c:504)]: Calling exit handler 16240
|
||||||
7
logs/eval/public/qwen3-0-6b@main/lambada.out
Normal file
7
logs/eval/public/qwen3-0-6b@main/lambada.out
Normal file
@ -0,0 +1,7 @@
|
|||||||
|
[RISE-CORE Msg(16052:140566486940672:libvgpu.c:901)]: Initializing.....
|
||||||
|
[RISE-CORE ERROR (pid:16052 thread=140566486940672 libvgpu.c:958)]: cuInit failed:100
|
||||||
|
02/04 11:11:52 - OpenCompass - INFO - Task [public/qwen3-0-6b@main/lambada]: {'accuracy': 0.038812342324859306}
|
||||||
|
02/04 11:11:52 - OpenCompass - INFO - time elapsed: 2.24s
|
||||||
|
/opt/conda/lib/python3.8/site-packages/fuzzywuzzy/fuzz.py:11: UserWarning: Using slow pure-python SequenceMatcher. Install python-Levenshtein to remove this warning
|
||||||
|
warnings.warn('Using slow pure-python SequenceMatcher. Install python-Levenshtein to remove this warning')
|
||||||
|
[RISE-CORE Msg(16052:140566486940672:multiprocess_memory_limit.c:504)]: Calling exit handler 16052
|
||||||
7
logs/eval/public/qwen3-0-6b@main/triviaqa.out
Normal file
7
logs/eval/public/qwen3-0-6b@main/triviaqa.out
Normal file
@ -0,0 +1,7 @@
|
|||||||
|
[RISE-CORE Msg(16237:140652034198528:libvgpu.c:901)]: Initializing.....
|
||||||
|
[RISE-CORE ERROR (pid:16237 thread=140652034198528 libvgpu.c:958)]: cuInit failed:100
|
||||||
|
02/04 11:12:04 - OpenCompass - INFO - Task [public/qwen3-0-6b@main/triviaqa]: {'score': 0.011316057485572028}
|
||||||
|
02/04 11:12:04 - OpenCompass - INFO - time elapsed: 3.91s
|
||||||
|
/opt/conda/lib/python3.8/site-packages/fuzzywuzzy/fuzz.py:11: UserWarning: Using slow pure-python SequenceMatcher. Install python-Levenshtein to remove this warning
|
||||||
|
warnings.warn('Using slow pure-python SequenceMatcher. Install python-Levenshtein to remove this warning')
|
||||||
|
[RISE-CORE Msg(16237:140652034198528:multiprocess_memory_limit.c:504)]: Calling exit handler 16237
|
||||||
File diff suppressed because one or more lines are too long
10
logs/infer/public/qwen3-0-6b@main/lambada_0.out
Normal file
10
logs/infer/public/qwen3-0-6b@main/lambada_0.out
Normal file
File diff suppressed because one or more lines are too long
10
logs/infer/public/qwen3-0-6b@main/lambada_1.out
Normal file
10
logs/infer/public/qwen3-0-6b@main/lambada_1.out
Normal file
File diff suppressed because one or more lines are too long
10
logs/infer/public/qwen3-0-6b@main/lambada_2.out
Normal file
10
logs/infer/public/qwen3-0-6b@main/lambada_2.out
Normal file
File diff suppressed because one or more lines are too long
10
logs/infer/public/qwen3-0-6b@main/triviaqa_0.out
Normal file
10
logs/infer/public/qwen3-0-6b@main/triviaqa_0.out
Normal file
File diff suppressed because one or more lines are too long
10
logs/infer/public/qwen3-0-6b@main/triviaqa_1.out
Normal file
10
logs/infer/public/qwen3-0-6b@main/triviaqa_1.out
Normal file
File diff suppressed because one or more lines are too long
10
logs/infer/public/qwen3-0-6b@main/triviaqa_2.out
Normal file
10
logs/infer/public/qwen3-0-6b@main/triviaqa_2.out
Normal file
File diff suppressed because one or more lines are too long
10
logs/infer/public/qwen3-0-6b@main/triviaqa_3.out
Normal file
10
logs/infer/public/qwen3-0-6b@main/triviaqa_3.out
Normal file
File diff suppressed because one or more lines are too long
10
logs/infer/public/qwen3-0-6b@main/triviaqa_4.out
Normal file
10
logs/infer/public/qwen3-0-6b@main/triviaqa_4.out
Normal file
File diff suppressed because one or more lines are too long
BIN
predictions/public/qwen3-0-6b@main/GaokaoBench_2010-2013_English_MCQs.json
(Stored with Git LFS)
Normal file
BIN
predictions/public/qwen3-0-6b@main/GaokaoBench_2010-2013_English_MCQs.json
(Stored with Git LFS)
Normal file
Binary file not shown.
BIN
predictions/public/qwen3-0-6b@main/lambada_0.json
(Stored with Git LFS)
Normal file
BIN
predictions/public/qwen3-0-6b@main/lambada_0.json
(Stored with Git LFS)
Normal file
Binary file not shown.
BIN
predictions/public/qwen3-0-6b@main/lambada_1.json
(Stored with Git LFS)
Normal file
BIN
predictions/public/qwen3-0-6b@main/lambada_1.json
(Stored with Git LFS)
Normal file
Binary file not shown.
BIN
predictions/public/qwen3-0-6b@main/lambada_2.json
(Stored with Git LFS)
Normal file
BIN
predictions/public/qwen3-0-6b@main/lambada_2.json
(Stored with Git LFS)
Normal file
Binary file not shown.
BIN
predictions/public/qwen3-0-6b@main/triviaqa_0.json
(Stored with Git LFS)
Normal file
BIN
predictions/public/qwen3-0-6b@main/triviaqa_0.json
(Stored with Git LFS)
Normal file
Binary file not shown.
BIN
predictions/public/qwen3-0-6b@main/triviaqa_1.json
(Stored with Git LFS)
Normal file
BIN
predictions/public/qwen3-0-6b@main/triviaqa_1.json
(Stored with Git LFS)
Normal file
Binary file not shown.
BIN
predictions/public/qwen3-0-6b@main/triviaqa_2.json
(Stored with Git LFS)
Normal file
BIN
predictions/public/qwen3-0-6b@main/triviaqa_2.json
(Stored with Git LFS)
Normal file
Binary file not shown.
BIN
predictions/public/qwen3-0-6b@main/triviaqa_3.json
(Stored with Git LFS)
Normal file
BIN
predictions/public/qwen3-0-6b@main/triviaqa_3.json
(Stored with Git LFS)
Normal file
Binary file not shown.
BIN
predictions/public/qwen3-0-6b@main/triviaqa_4.json
(Stored with Git LFS)
Normal file
BIN
predictions/public/qwen3-0-6b@main/triviaqa_4.json
(Stored with Git LFS)
Normal file
Binary file not shown.
BIN
results/public/qwen3-0-6b@main/GaokaoBench_2010-2013_English_MCQs.json
(Stored with Git LFS)
Normal file
BIN
results/public/qwen3-0-6b@main/GaokaoBench_2010-2013_English_MCQs.json
(Stored with Git LFS)
Normal file
Binary file not shown.
BIN
results/public/qwen3-0-6b@main/lambada.json
(Stored with Git LFS)
Normal file
BIN
results/public/qwen3-0-6b@main/lambada.json
(Stored with Git LFS)
Normal file
Binary file not shown.
BIN
results/public/qwen3-0-6b@main/triviaqa.json
(Stored with Git LFS)
Normal file
BIN
results/public/qwen3-0-6b@main/triviaqa.json
(Stored with Git LFS)
Normal file
Binary file not shown.
87
summary/summary_20260204_103117.csv
Normal file
87
summary/summary_20260204_103117.csv
Normal file
@ -0,0 +1,87 @@
|
|||||||
|
dataset,version,metric,mode,public/qwen3-0-6b@main
|
||||||
|
--------- 考试 Exam ---------,-,-,-,-
|
||||||
|
ceval,-,-,-,-
|
||||||
|
agieval,-,-,-,-
|
||||||
|
mmlu,-,-,-,-
|
||||||
|
GaokaoBench,-,-,-,-
|
||||||
|
ARC-c,-,-,-,-
|
||||||
|
--------- 语言 Language ---------,-,-,-,-
|
||||||
|
WiC,-,-,-,-
|
||||||
|
summedits,-,-,-,-
|
||||||
|
chid-dev,-,-,-,-
|
||||||
|
afqmc-dev,-,-,-,-
|
||||||
|
bustm-dev,-,-,-,-
|
||||||
|
cluewsc-dev,-,-,-,-
|
||||||
|
WSC,-,-,-,-
|
||||||
|
winogrande,-,-,-,-
|
||||||
|
flores_100,-,-,-,-
|
||||||
|
--------- 知识 Knowledge ---------,-,-,-,-
|
||||||
|
BoolQ,-,-,-,-
|
||||||
|
commonsense_qa,-,-,-,-
|
||||||
|
nq,-,-,-,-
|
||||||
|
triviaqa,2121ce,score,gen,0.01
|
||||||
|
--------- 推理 Reasoning ---------,-,-,-,-
|
||||||
|
cmnli,-,-,-,-
|
||||||
|
ocnli,-,-,-,-
|
||||||
|
ocnli_fc-dev,-,-,-,-
|
||||||
|
AX_b,-,-,-,-
|
||||||
|
AX_g,-,-,-,-
|
||||||
|
CB,-,-,-,-
|
||||||
|
RTE,-,-,-,-
|
||||||
|
story_cloze,-,-,-,-
|
||||||
|
COPA,-,-,-,-
|
||||||
|
ReCoRD,-,-,-,-
|
||||||
|
hellaswag,-,-,-,-
|
||||||
|
piqa,-,-,-,-
|
||||||
|
siqa,-,-,-,-
|
||||||
|
strategyqa,-,-,-,-
|
||||||
|
math,-,-,-,-
|
||||||
|
gsm8k,-,-,-,-
|
||||||
|
TheoremQA,-,-,-,-
|
||||||
|
openai_humaneval,-,-,-,-
|
||||||
|
mbpp,-,-,-,-
|
||||||
|
cmmlu,-,-,-,-
|
||||||
|
bbh,-,-,-,-
|
||||||
|
--------- 理解 Understanding ---------,-,-,-,-
|
||||||
|
C3,-,-,-,-
|
||||||
|
CMRC_dev,-,-,-,-
|
||||||
|
DRCD_dev,-,-,-,-
|
||||||
|
MultiRC,-,-,-,-
|
||||||
|
race-middle,-,-,-,-
|
||||||
|
race-high,-,-,-,-
|
||||||
|
openbookqa_fact,-,-,-,-
|
||||||
|
csl_dev,-,-,-,-
|
||||||
|
lcsts,-,-,-,-
|
||||||
|
Xsum,-,-,-,-
|
||||||
|
eprstmt-dev,-,-,-,-
|
||||||
|
lambada,217e11,accuracy,gen,0.04
|
||||||
|
tnews-dev,-,-,-,-
|
||||||
|
--------- 安全 Safety ---------,-,-,-,-
|
||||||
|
crows_pairs,-,-,-,-
|
||||||
|
--------- LEval Exact Match (Acc) ---------,-,-,-,-
|
||||||
|
LEval_coursera,-,-,-,-
|
||||||
|
LEval_gsm100,-,-,-,-
|
||||||
|
LEval_quality,-,-,-,-
|
||||||
|
LEval_tpo,-,-,-,-
|
||||||
|
LEval_topic_retrieval,-,-,-,-
|
||||||
|
--------- LEval Gen (ROUGE) ---------,-,-,-,-
|
||||||
|
LEval_financialqa,-,-,-,-
|
||||||
|
LEval_gov_report_summ,-,-,-,-
|
||||||
|
LEval_legal_contract_qa,-,-,-,-
|
||||||
|
LEval_meeting_summ,-,-,-,-
|
||||||
|
LEval_multidocqa,-,-,-,-
|
||||||
|
LEval_narrativeqa,-,-,-,-
|
||||||
|
LEval_nq,-,-,-,-
|
||||||
|
LEval_news_summ,-,-,-,-
|
||||||
|
LEval_paper_assistant,-,-,-,-
|
||||||
|
LEval_patent_summ,-,-,-,-
|
||||||
|
LEval_review_summ,-,-,-,-
|
||||||
|
LEval_scientificqa,-,-,-,-
|
||||||
|
LEval_tvshow_summ--------- 长文本 LongBench ---------,-,-,-,-
|
||||||
|
longbench_lsht,-,-,-,-
|
||||||
|
longbench_vcsum,-,-,-,-
|
||||||
|
longbench_dureader,-,-,-,-
|
||||||
|
longbench_multifieldqa_zh,-,-,-,-
|
||||||
|
longbench_passage_retrieval_zh,-,-,-,-
|
||||||
|
--------- 单选 自定义数据 ---------,-,-,-,-
|
||||||
|
SageBench-exam,-,-,-,-
|
||||||
|
197
summary/summary_20260204_103117.txt
Normal file
197
summary/summary_20260204_103117.txt
Normal file
@ -0,0 +1,197 @@
|
|||||||
|
20260204_103117
|
||||||
|
tabulate format
|
||||||
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
dataset version metric mode public/qwen3-0-6b@main
|
||||||
|
----------------------------------------------------- --------- -------- ------ ------------------------
|
||||||
|
--------- 考试 Exam --------- - - - -
|
||||||
|
ceval - - - -
|
||||||
|
agieval - - - -
|
||||||
|
mmlu - - - -
|
||||||
|
GaokaoBench - - - -
|
||||||
|
ARC-c - - - -
|
||||||
|
--------- 语言 Language --------- - - - -
|
||||||
|
WiC - - - -
|
||||||
|
summedits - - - -
|
||||||
|
chid-dev - - - -
|
||||||
|
afqmc-dev - - - -
|
||||||
|
bustm-dev - - - -
|
||||||
|
cluewsc-dev - - - -
|
||||||
|
WSC - - - -
|
||||||
|
winogrande - - - -
|
||||||
|
flores_100 - - - -
|
||||||
|
--------- 知识 Knowledge --------- - - - -
|
||||||
|
BoolQ - - - -
|
||||||
|
commonsense_qa - - - -
|
||||||
|
nq - - - -
|
||||||
|
triviaqa 2121ce score gen 0.01
|
||||||
|
--------- 推理 Reasoning --------- - - - -
|
||||||
|
cmnli - - - -
|
||||||
|
ocnli - - - -
|
||||||
|
ocnli_fc-dev - - - -
|
||||||
|
AX_b - - - -
|
||||||
|
AX_g - - - -
|
||||||
|
CB - - - -
|
||||||
|
RTE - - - -
|
||||||
|
story_cloze - - - -
|
||||||
|
COPA - - - -
|
||||||
|
ReCoRD - - - -
|
||||||
|
hellaswag - - - -
|
||||||
|
piqa - - - -
|
||||||
|
siqa - - - -
|
||||||
|
strategyqa - - - -
|
||||||
|
math - - - -
|
||||||
|
gsm8k - - - -
|
||||||
|
TheoremQA - - - -
|
||||||
|
openai_humaneval - - - -
|
||||||
|
mbpp - - - -
|
||||||
|
cmmlu - - - -
|
||||||
|
bbh - - - -
|
||||||
|
--------- 理解 Understanding --------- - - - -
|
||||||
|
C3 - - - -
|
||||||
|
CMRC_dev - - - -
|
||||||
|
DRCD_dev - - - -
|
||||||
|
MultiRC - - - -
|
||||||
|
race-middle - - - -
|
||||||
|
race-high - - - -
|
||||||
|
openbookqa_fact - - - -
|
||||||
|
csl_dev - - - -
|
||||||
|
lcsts - - - -
|
||||||
|
Xsum - - - -
|
||||||
|
eprstmt-dev - - - -
|
||||||
|
lambada 217e11 accuracy gen 0.04
|
||||||
|
tnews-dev - - - -
|
||||||
|
--------- 安全 Safety --------- - - - -
|
||||||
|
crows_pairs - - - -
|
||||||
|
--------- LEval Exact Match (Acc) --------- - - - -
|
||||||
|
LEval_coursera - - - -
|
||||||
|
LEval_gsm100 - - - -
|
||||||
|
LEval_quality - - - -
|
||||||
|
LEval_tpo - - - -
|
||||||
|
LEval_topic_retrieval - - - -
|
||||||
|
--------- LEval Gen (ROUGE) --------- - - - -
|
||||||
|
LEval_financialqa - - - -
|
||||||
|
LEval_gov_report_summ - - - -
|
||||||
|
LEval_legal_contract_qa - - - -
|
||||||
|
LEval_meeting_summ - - - -
|
||||||
|
LEval_multidocqa - - - -
|
||||||
|
LEval_narrativeqa - - - -
|
||||||
|
LEval_nq - - - -
|
||||||
|
LEval_news_summ - - - -
|
||||||
|
LEval_paper_assistant - - - -
|
||||||
|
LEval_patent_summ - - - -
|
||||||
|
LEval_review_summ - - - -
|
||||||
|
LEval_scientificqa - - - -
|
||||||
|
LEval_tvshow_summ--------- 长文本 LongBench --------- - - - -
|
||||||
|
longbench_lsht - - - -
|
||||||
|
longbench_vcsum - - - -
|
||||||
|
longbench_dureader - - - -
|
||||||
|
longbench_multifieldqa_zh - - - -
|
||||||
|
longbench_passage_retrieval_zh - - - -
|
||||||
|
--------- 单选 自定义数据 --------- - - - -
|
||||||
|
SageBench-exam - - - -
|
||||||
|
$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$
|
||||||
|
|
||||||
|
-------------------------------------------------------------------------------------------------------------------------------- THIS IS A DIVIDER --------------------------------------------------------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
csv format
|
||||||
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
dataset,version,metric,mode,public/qwen3-0-6b@main
|
||||||
|
--------- 考试 Exam ---------,-,-,-,-
|
||||||
|
ceval,-,-,-,-
|
||||||
|
agieval,-,-,-,-
|
||||||
|
mmlu,-,-,-,-
|
||||||
|
GaokaoBench,-,-,-,-
|
||||||
|
ARC-c,-,-,-,-
|
||||||
|
--------- 语言 Language ---------,-,-,-,-
|
||||||
|
WiC,-,-,-,-
|
||||||
|
summedits,-,-,-,-
|
||||||
|
chid-dev,-,-,-,-
|
||||||
|
afqmc-dev,-,-,-,-
|
||||||
|
bustm-dev,-,-,-,-
|
||||||
|
cluewsc-dev,-,-,-,-
|
||||||
|
WSC,-,-,-,-
|
||||||
|
winogrande,-,-,-,-
|
||||||
|
flores_100,-,-,-,-
|
||||||
|
--------- 知识 Knowledge ---------,-,-,-,-
|
||||||
|
BoolQ,-,-,-,-
|
||||||
|
commonsense_qa,-,-,-,-
|
||||||
|
nq,-,-,-,-
|
||||||
|
triviaqa,2121ce,score,gen,0.01
|
||||||
|
--------- 推理 Reasoning ---------,-,-,-,-
|
||||||
|
cmnli,-,-,-,-
|
||||||
|
ocnli,-,-,-,-
|
||||||
|
ocnli_fc-dev,-,-,-,-
|
||||||
|
AX_b,-,-,-,-
|
||||||
|
AX_g,-,-,-,-
|
||||||
|
CB,-,-,-,-
|
||||||
|
RTE,-,-,-,-
|
||||||
|
story_cloze,-,-,-,-
|
||||||
|
COPA,-,-,-,-
|
||||||
|
ReCoRD,-,-,-,-
|
||||||
|
hellaswag,-,-,-,-
|
||||||
|
piqa,-,-,-,-
|
||||||
|
siqa,-,-,-,-
|
||||||
|
strategyqa,-,-,-,-
|
||||||
|
math,-,-,-,-
|
||||||
|
gsm8k,-,-,-,-
|
||||||
|
TheoremQA,-,-,-,-
|
||||||
|
openai_humaneval,-,-,-,-
|
||||||
|
mbpp,-,-,-,-
|
||||||
|
cmmlu,-,-,-,-
|
||||||
|
bbh,-,-,-,-
|
||||||
|
--------- 理解 Understanding ---------,-,-,-,-
|
||||||
|
C3,-,-,-,-
|
||||||
|
CMRC_dev,-,-,-,-
|
||||||
|
DRCD_dev,-,-,-,-
|
||||||
|
MultiRC,-,-,-,-
|
||||||
|
race-middle,-,-,-,-
|
||||||
|
race-high,-,-,-,-
|
||||||
|
openbookqa_fact,-,-,-,-
|
||||||
|
csl_dev,-,-,-,-
|
||||||
|
lcsts,-,-,-,-
|
||||||
|
Xsum,-,-,-,-
|
||||||
|
eprstmt-dev,-,-,-,-
|
||||||
|
lambada,217e11,accuracy,gen,0.04
|
||||||
|
tnews-dev,-,-,-,-
|
||||||
|
--------- 安全 Safety ---------,-,-,-,-
|
||||||
|
crows_pairs,-,-,-,-
|
||||||
|
--------- LEval Exact Match (Acc) ---------,-,-,-,-
|
||||||
|
LEval_coursera,-,-,-,-
|
||||||
|
LEval_gsm100,-,-,-,-
|
||||||
|
LEval_quality,-,-,-,-
|
||||||
|
LEval_tpo,-,-,-,-
|
||||||
|
LEval_topic_retrieval,-,-,-,-
|
||||||
|
--------- LEval Gen (ROUGE) ---------,-,-,-,-
|
||||||
|
LEval_financialqa,-,-,-,-
|
||||||
|
LEval_gov_report_summ,-,-,-,-
|
||||||
|
LEval_legal_contract_qa,-,-,-,-
|
||||||
|
LEval_meeting_summ,-,-,-,-
|
||||||
|
LEval_multidocqa,-,-,-,-
|
||||||
|
LEval_narrativeqa,-,-,-,-
|
||||||
|
LEval_nq,-,-,-,-
|
||||||
|
LEval_news_summ,-,-,-,-
|
||||||
|
LEval_paper_assistant,-,-,-,-
|
||||||
|
LEval_patent_summ,-,-,-,-
|
||||||
|
LEval_review_summ,-,-,-,-
|
||||||
|
LEval_scientificqa,-,-,-,-
|
||||||
|
LEval_tvshow_summ--------- 长文本 LongBench ---------,-,-,-,-
|
||||||
|
longbench_lsht,-,-,-,-
|
||||||
|
longbench_vcsum,-,-,-,-
|
||||||
|
longbench_dureader,-,-,-,-
|
||||||
|
longbench_multifieldqa_zh,-,-,-,-
|
||||||
|
longbench_passage_retrieval_zh,-,-,-,-
|
||||||
|
--------- 单选 自定义数据 ---------,-,-,-,-
|
||||||
|
SageBench-exam,-,-,-,-
|
||||||
|
$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$
|
||||||
|
|
||||||
|
-------------------------------------------------------------------------------------------------------------------------------- THIS IS A DIVIDER --------------------------------------------------------------------------------------------------------------------------------
|
||||||
|
|
||||||
|
raw format
|
||||||
|
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
|
||||||
|
-------------------------------
|
||||||
|
Model: public/qwen3-0-6b@main
|
||||||
|
GaokaoBench_2010-2013_English_MCQs: {'score': 40.0}
|
||||||
|
lambada: {'accuracy': 0.038812342324859306}
|
||||||
|
triviaqa: {'score': 0.011316057485572028}
|
||||||
|
GaokaoBench: {'error': "missing datasets: {'GaokaoBench_2010-2022_Chinese_Modern_Lit', 'GaokaoBench_2010-2022_Chinese_Lang_and_Usage_MCQs', 'GaokaoBench_2010-2022_Chemistry_MCQs', 'GaokaoBench_2010-2022_Physics_MCQs', 'GaokaoBench_2010-2022_Math_II_MCQs', 'GaokaoBench_2010-2022_Geography_MCQs', 'GaokaoBench_2010-2022_English_Fill_in_Blanks', 'GaokaoBench_2012-2022_English_Cloze_Test', 'GaokaoBench_2010-2022_English_Reading_Comp', 'GaokaoBench_2010-2022_History_MCQs', 'GaokaoBench_2010-2022_Political_Science_MCQs', 'GaokaoBench_2010-2022_Math_I_MCQs', 'GaokaoBench_2010-2022_Biology_MCQs'}"}
|
||||||
|
$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$$
|
||||||
Loading…
Reference in New Issue
Block a user