File size: 10,768 Bytes
e150b93
3e2fdcf
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
	eval_name	Precision	Type	T	Weight type	Architecture	Model	Average ⬆️	Hub License	#Params (B)	Model sha	Hub ❤️	Available on the hub	Code	Data	AMI 2020 Agg	AMI 2020 Miso	ARC-C	Belebele	GeNTE Neutralizing	HaSpeeDe2 HS	HaSpeeDe2 Stereo	HateCheck	HONEST	IronITA Irony	IronITA Sarcasm	ItaCoLA	News Sum	SENTIPOLC	SQuAD it	TruthfulQA	XCOPA
0	swap-uniba_LLaMAntino-3-ANITA-8B-Inst-DPO-ITA_?	?		?	Unknown	LlamaForCausalLM	"<a target=""_blank"" href=""https://huggingface.co/swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA</a>"	59.925692200303324				0	True	?	?	49.61934617107031	73.58604698054239	56.484641638225256	83.55555555555556	33.8255033557047	72.24399819126907	61.627116844508144	80.51511613552358	100.0	67.79529918401192	46.19514665929917	0.24261234404280246	33.783978293075634	46.49499761664646	71.27317142821833	68.09423700746308	73.4
9	mistralai_Mistral-7B-Instruct-v0.2_?	?		?	Unknown	MistralForCausalLM	"<a target=""_blank"" href=""https://huggingface.co/mistralai/Mistral-7B-Instruct-v0.2"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">mistralai/Mistral-7B-Instruct-v0.2</a>"	57.57154925481929				0	True	?	?	61.95096430524839	66.42194008585093	44.36860068259386	67.22222222222223	29.12751677852349	71.07491292799637	67.27017961567233	78.40873056250285	100.0	59.16469471738617	55.53851376330874	0.27708420363666786	36.377962201593874	50.02052664310759	68.04841543730598	59.24407318497844	64.2
2	mii-community_zefiro-7b-dpo-ITA_?	?		?	Unknown	MistralForCausalLM	"<a target=""_blank"" href=""https://huggingface.co/mii-community/zefiro-7b-dpo-ITA"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">mii-community/zefiro-7b-dpo-ITA</a>"	55.96825697198048				0	True	?	?	59.97920997920998	66.14027143881808	44.19795221843004	65.88888888888889	29.798657718120808	66.93068606112085	61.46209894750329	82.83622905315102	100.0	58.523449206965395	54.918191698733956	0.22337556862808253	35.66642647158017	38.80971929318383	74.34293876621986	43.34227321311386	68.4
7	meta-llama_Meta-Llama-3-8B_?	?		?	Unknown	LlamaForCausalLM	"<a target=""_blank"" href=""https://huggingface.co/meta-llama/Meta-Llama-3-8B"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">meta-llama/Meta-Llama-3-8B</a>"	55.933099551030125				0	True	?	?	60.02710027100271	63.14678395603251	40.529010238907844	76.0	29.53020134228188	65.30297764359561	59.541073390095356	79.81131536880565	100.0	57.31801541230962	56.750548188367965	0.2786244415689118	32.93607461627173	39.93136214294286	76.49082768675667	42.06877766857276	71.2
10	mii-community_zefiro-7b-base-ITA_?	?		?	Unknown	MistralForCausalLM	"<a target=""_blank"" href=""https://huggingface.co/mii-community/zefiro-7b-base-ITA"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">mii-community/zefiro-7b-base-ITA</a>"	55.286768709834995				0	True	?	?	60.14362403797995	64.54082375784897	40.955631399317404	58.55555555555556	28.456375838926174	66.12858980217781	59.74063711314884	82.46753086246828	100.0	59.05311714498798	57.8863223808017	0.09963712635854956	34.19887652648641	39.18986054178559	75.6692177776856	46.18926820166605	66.60000000000001
15	mii-community_zefiro-7b-sft-ITA_?	?		?	Unknown	MistralForCausalLM	"<a target=""_blank"" href=""https://huggingface.co/mii-community/zefiro-7b-sft-ITA"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">mii-community/zefiro-7b-sft-ITA</a>"	55.135348610310785				0	True	?	?	60.458679319889285	63.51256529535591	42.32081911262799	67.77777777777779	27.248322147651006	65.72752014372092	60.158604473839915	83.05031763559394	100.0	52.69566548195397	51.630329924754	0.08940878967203518	34.80608014621687	43.75098014181036	74.55382319645513	42.52003278796414	67.0
1	mistralai_Mistral-7B-v0.1_?	?		?	Unknown	MistralForCausalLM	"<a target=""_blank"" href=""https://huggingface.co/mistralai/Mistral-7B-v0.1"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">mistralai/Mistral-7B-v0.1</a>"	54.550973703693096				0	True	?	?	60.52050697114497	63.66158365032981	41.21160409556314	65.66666666666666	29.53020134228188	60.38816689466484	57.907599364752336	80.59264657366079	100.0	55.23299236027556	55.67900219124808	0.131895692851752	34.09475870496535	38.87141003943634	75.08500650762954	43.19251190731156	65.60000000000001
14	swap-uniba_LLaMAntino-2-chat-13b-hf-ITA_?	?		?	Unknown	LlamaForCausalLM	"<a target=""_blank"" href=""https://huggingface.co/swap-uniba/LLaMAntino-2-chat-13b-hf-ITA"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">swap-uniba/LLaMAntino-2-chat-13b-hf-ITA</a>"	53.88562700961127				0	True	?	?	61.41230947327803	64.77739009492042	39.07849829351536	60.44444444444444	25.503355704697988	67.1548291501024	59.101414060364085	81.83763297921335	100.0	57.92048929663609	52.2777996043644	0.1015435288181161	23.81691473597593	34.69232896418751	73.10003377486571	44.43667505800782	70.39999999999999
6	swap-uniba_LLaMAntino-2-13b-hf-ITA_?	?		?	Unknown	LlamaForCausalLM	"<a target=""_blank"" href=""https://huggingface.co/swap-uniba/LLaMAntino-2-13b-hf-ITA"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">swap-uniba/LLaMAntino-2-13b-hf-ITA</a>"	51.26001015437534				0	True	?	?	56.79723502304148	60.93495016444478	38.56655290102389	52.33333333333333	24.697986577181208	57.1976786986929	54.2447910290625	68.16391542846057	100.0	56.51605280366516	51.571111501558086	0.16387751408972254	23.495330157527007	38.60258050721683	74.20709928774112	42.12767769734222	71.8
13	meta-llama_Llama-2-13b-hf_?	?		?	Unknown	LlamaForCausalLM	"<a target=""_blank"" href=""https://huggingface.co/meta-llama/Llama-2-13b-hf"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">meta-llama/Llama-2-13b-hf</a>"	51.16172039685661				0	True	?	?	53.24565637065637	59.32319654843206	39.93174061433447	49.666666666666664	24.295302013422816	54.13600451447075	54.88702987697715	74.1483219663718	100.0	50.34584608393744	49.636673785442774	0.11758183179468357	35.09699883531247	37.37259554778931	75.22840229480128	42.91722979615231	69.39999999999999
5	g8a9_tweety-mistral-7b_?	?		?	Unknown	MistralForCausalLM	"<a target=""_blank"" href=""https://huggingface.co/g8a9/tweety-mistral-7b"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">g8a9/tweety-mistral-7b</a>"	48.37800669811661				0	True	?	?	56.17170479302832	56.423255312264054	37.96928327645051	49.666666666666664	27.91946308724832	53.70259637851317	53.57434872305199	64.41588573083048	100.0	50.21506876304183	49.42973129711966	0.11006633622278786	18.81035591897043	28.46426204947685	64.39794432633592	37.75548120876122	73.4
4	meta-llama_Llama-2-7b-hf_?	?		?	Unknown	LlamaForCausalLM	"<a target=""_blank"" href=""https://huggingface.co/meta-llama/Llama-2-7b-hf"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">meta-llama/Llama-2-7b-hf</a>"	47.26821759114118				0	True	?	?	50.26836062232489	57.089775606014214	35.153583617747444	36.11111111111111	25.100671140939596	49.33536331841416	51.73318260900284	67.35406316275402	100.0	47.63910390674802	48.347086153434084	0.036528464070504335	33.756452251726735	27.82288694076669	68.6449557225095	39.16657442183614	66.0
8	swap-uniba_LLaMAntino-2-7b-hf-ITA_?	?		?	Unknown	LlamaForCausalLM	"<a target=""_blank"" href=""https://huggingface.co/swap-uniba/LLaMAntino-2-7b-hf-ITA"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">swap-uniba/LLaMAntino-2-7b-hf-ITA</a>"	45.2518617559276				0	True	?	?	51.11111111111111	53.267951636107355	33.70307167235495	34.66666666666667	24.295302013422816	45.514286626950536	47.59019966407009	60.855425171736485	100.0	47.55193616643805	46.04838972288254	0.043130721156949686	24.582547279426233	22.260015178994326	69.30864535653794	40.48297086291322	68.0
12	sapienzanlp_Minerva-3B-base-v1.0_?	?		?	Unknown	MistralForCausalLM	"<a target=""_blank"" href=""https://huggingface.co/sapienzanlp/Minerva-3B-base-v1.0"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">sapienzanlp/Minerva-3B-base-v1.0</a>"	41.485187669928465				0	True	?	?	49.875480140137604	52.15633707230505	30.97269624573379	24.333333333333336	23.08724832214765	48.93622623624203	45.71528801169143	47.43110547988597	100.0	43.13118956315911	45.77114427860697	-0.015363788820154219	21.8700732759084	23.020245154283693	42.99174436502196	37.371442699146954	68.60000000000001
3	swap-uniba_LLaMAntino-2-chat-7b-hf-ITA_?	?		?	Unknown	LlamaForCausalLM	"<a target=""_blank"" href=""https://huggingface.co/swap-uniba/LLaMAntino-2-chat-7b-hf-ITA"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">swap-uniba/LLaMAntino-2-chat-7b-hf-ITA</a>"	38.997012161113425				0	True	?	?	47.32809806550469	43.776841477788466	29.180887372013654	28.111111111111107	23.48993288590604	41.57668822526659	41.556830771361305	44.984357634264406	100.0	41.716872329343005	43.53102430893341	-0.02574637563194932	8.269309204888462	9.339380225529704	58.43272201840739	39.880897484241935	61.8
16	sapienzanlp_Minerva-1B-base-v1.0_?	?		?	Unknown	MistralForCausalLM	"<a target=""_blank"" href=""https://huggingface.co/sapienzanlp/Minerva-1B-base-v1.0"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">sapienzanlp/Minerva-1B-base-v1.0</a>"	38.906733116823304				0	True	?	?	50.76172656624852	53.84641914146224	24.573378839590443	22.666666666666664	26.57718120805369	48.25128927047713	44.581537708222804	50.10425395808837	100.0	46.49541549308013	45.46046920890855	0.022249590030925144	14.27287574762189	16.571464690513597	17.48160254077023	39.747932356260876	60.0
11	sapienzanlp_Minerva-350M-base-v1.0_?	?		?	Unknown	MistralForCausalLM	"<a target=""_blank"" href=""https://huggingface.co/sapienzanlp/Minerva-350M-base-v1.0"" style=""color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;"">sapienzanlp/Minerva-350M-base-v1.0</a>"	36.95204565967993				0	True	?	?	45.17543859649123	35.72145622912868	24.40273037542662	22.88888888888889	52.75167785234899	41.92832319168979	40.67042217927179	46.277755136438564	100.0	36.23277134884009	43.223117993157416	-0.036868413829916326	10.308018221966565	23.388373345290127	4.903980027793706	43.7486912416563	56.599999999999994