0daa7a78d6ba552d3ca93170d9f7455409a24a17
Transformer decoder models (CLM) transformer huggingface-transformers transformer-decoder clm

 

challenge
GLUE-LM-GAP
submitter
kaczla
submitted
2023-02-13 09:07:21.810036 UTC
original repo
https://gitlab.com/kaczla/glue-lm-gap.git / branch decoder
publicly available at
git://gonito.net/glue-lm-gap / branch submission-07997
browsable at
https://gonito.net/gitlist/glue-lm-gap.git/submission-07997/
clone by
git clone --single-branch git://gonito.net/glue-lm-gap -b submission-07997
# model_name dev-0 PerplexityHashed test-A PerplexityHashed
93 PolishGPT-2-small 79627.908898 79278.556763
88 PolishGPT-2-medium 59754.664598 58446.342067
85 PolishGPT-2-large 50521.897871 47767.398812
84 GPT-fr-base 46956.696551 46362.459344
83 GPT-fr-small 38383.860649 38199.695407
72 DistilGPT-2 8885.673813 9714.724445
70 BioGPT 10345.246977 9377.074086
68 Pythia-70M 7424.531362 7329.966876
67 Pythia-160M 5019.867530 4845.613045
66 GPT-2-base 4732.154178 4700.318867
62 GPT-Neo-125M 4711.535675 4373.950153
60 Pythia-410M 3894.413763 3685.007192
54 GPT-2-medium 2591.721115 2250.943823
51 OPT-125M 2560.284962 2021.615293
49 OPT-350M 2257.973076 1767.228754
Parameter Value
depth 1
top_k 15