Dataset Viewer
Auto-converted to Parquet Duplicate
musicmc
float64
0.08
0.41
mmluproru
float64
0.1
0.37
lawmc
float64
0.11
0.75
model
stringlengths
13
55
moviesmc
float64
0.07
0.75
booksmc
float64
0.08
0.53
model_dtype
stringclasses
2 values
ppl
int64
0
0
0.306383
0.276166
0.623069
BlackSamorez/TechxGenus/Mistral-Large-Instruct-2407-AWQ
0.715278
0.446058
torch.float16
0
0.165957
0.168912
0.319258
LakoMoor/Vikhr-2.5-7b-it
0.328704
0.209544
torch.float16
0
0.27234
null
0.485067
Nexusflow/Starling-LM-7B-beta
0.386574
0.307054
torch.float16
0
0.093617
0.102073
0.114315
NousResearch/Llama-2-7b-hf
0.071759
0.107884
torch.float16
0
0.319149
0.219171
0.607621
RefalMachine/IlyaGusev_saiga_llama3_8b
0.474537
0.360996
torch.bfloat16
0
0.32766
0.260104
0.553038
RefalMachine/Qwen2.5-32B-instruct
0.62037
0.414938
torch.bfloat16
0
0.13617
0.186528
0.235839
RefalMachine/Qwen2.5-3B-Instruct
0.210648
0.172199
torch.bfloat16
0
0.13617
0.186528
0.235839
RefalMachine/Qwen2.5-7B-instruct
0.210648
0.172199
torch.bfloat16
0
0.351064
0.323834
0.698249
RefalMachine/RuadaptQwen-32B-instruct
0.643519
0.487552
torch.bfloat16
0
0.325532
0.289119
0.687951
RefalMachine/RuadaptQwen2.5-14B-Instruct-v1
0.613426
0.46888
torch.float16
0
0.410638
0.370984
0.744593
RefalMachine/RuadaptQwen2.5-32B-Pro
0.6875
0.526971
torch.bfloat16
0
0.261702
0.221762
0.542739
RefalMachine/RuadaptQwen2.5-3B-instruct
0.321759
0.317427
torch.bfloat16
0
0.348936
0.264249
0.645726
RefalMachine/RuadaptQwen2.5-7B-Lite-v1
0.550926
0.462656
torch.float16
0
0.291489
0.222798
0.624099
RefalMachine/openchat-3.5-0106
0.453704
0.390041
torch.bfloat16
0
0.310638
0.22487
0.61689
RefalMachine/ruadapt_llama3_8b_instruct_extended_led_ft
0.474537
0.392116
torch.bfloat16
0
0.208511
null
0.471679
Salesforce/LLaMA-3-8B-SFR-Iterative-DPO-R
0.305556
0.261411
torch.float16
0
0.268085
0.201036
0.53862
Vikhrmodels/it-5.2-fp16-cp
0.453704
0.307054
torch.float16
0
0.302128
null
0.544799
alexwortega/saiga_submit
0.395833
0.338174
torch.bfloat16
0
0.321277
0.251813
0.728115
apsys/GigaChatLite-chat-template
0.696759
0.446058
torch.float16
0
0.346809
0.301554
0.750772
apsys/GigaChatMax-chat-template
0.74537
0.520747
torch.float16
0
0.312766
0.262176
0.728115
apsys/GigaChatPro-chat-template
0.641204
0.443983
torch.float16
0
0.280851
0.179793
0.532441
apsys/T-lite-instruct-0.1
0.469907
0.3361
torch.float16
0
0.076596
0.123316
0.188465
apsys/llama3_8b_it
0.083333
0.080913
torch.float16
0
0.076596
0.123834
0.23172
apsys/llama3_overfitted
0.099537
0.087137
torch.float16
0
0.280851
0.179793
0.532441
apsys/tlite-it-0.1
0.469907
0.3361
torch.float16
0
0.287234
null
0.506694
vikhr-52-7b-chat-hf/apsys
0.483796
0.307054
torch.float16
0
0.280851
0.188083
0.642636
apsys/vikhr-it-5.4-fp16-orpo-v2
0.469907
0.334025
torch.float16
0
0.208511
null
0.426365
cohere/aya-8b
0.328704
0.242739
torch.float16
0
0.255319
0.262176
0.581874
google/gemma-2-9b
0.50463
0.3361
torch.float16
0
0.259574
0.193782
0.518023
lightblue/suzume-llama-3-8B-multilingual
0.328704
0.29668
torch.float16
0
0.293617
null
0.534501
RefalMachine/llama3 ushanka
0.351852
0.325726
torch.bfloat16
0
0.282979
null
0.54068
microsoft/Phi-3-medium-4k-instruct
0.428241
0.381743
torch.float16
0
0.293617
null
0.480947
apsys/saiga_3_8b
0.340278
0.311203
torch.float16
0
0.302128
null
0.544799
IlyaGusev/saiga_llama3_8b
0.395833
0.338174
torch.bfloat16
0
0.251064
null
0.487127
apsys/vikhr-52-7b
0.421296
0.311203
torch.float16
0
0.244681
null
0.478888
apsys/vikhr-53-7b-32k
0.405093
0.304979
torch.float16
0
README.md exists but content is empty.
Downloads last month
801

Spaces using Vikhrmodels/s-shlepa-metainfo 2