type
stringclasses 1
value | id
stringlengths 5
122
| num_branches
int64 1
1.76k
| branches
listlengths 1
1.76k
| main_branch_size
int64 0
32,943B
|
---|---|---|---|---|
model
|
ashnadua01/model_muril_small_noise
| 1 |
[
"main"
] | 957,531,409 |
model
|
gbueno86/Meta-Llama-3-120b-LumiLumimaid
| 1 |
[
"main"
] | 243,795,340,360 |
model
|
liilliliiiliil/VALORANT_SOVA
| 1 |
[
"main"
] | 115,055,119 |
model
|
mradermacher/Rhea-125-v0.5-i1-GGUF
| 1 |
[
"main"
] | 1,207,383,908,874 |
model
|
sel639/llava-1.5-7b-hf-ft-bill-epoch-50-continue
| 1 |
[
"main"
] | 1,519 |
model
|
NicolasAdrian07/Lockjaw_TRTF
| 1 |
[
"main"
] | 54,609,143 |
model
|
fth66/2nd_sub
| 1 |
[
"main"
] | 328,031 |
model
|
GitBag/rebel_nectar_1715368613
| 12 |
[
"step2201",
"step1201",
"step1401",
"step1801",
"step1001",
"step2001",
"step801",
"step1601",
"step201",
"step601",
"step401",
"main"
] | 1,519 |
model
|
AwsafAlam/flan-t5-large-code-review-lora-may-12-2024
| 1 |
[
"main"
] | 75,550,890 |
model
|
GoldoBasic/Plory_RVC
| 1 |
[
"main"
] | 62,227,240 |
model
|
alexandra-udrescu/qa_toolformer
| 1 |
[
"main"
] | 134,266,255 |
model
|
Felladrin/gguf-sharded-Phi-3-mini-4k-instruct-iMat
| 1 |
[
"main"
] | 2,815,287,511 |
model
|
saeidebbei/Azkoon
| 1 |
[
"main"
] | 1,519 |
model
|
azsxscdvfb/VetMedGPT-1B-V0.2
| 1 |
[
"main"
] | 8,038,858,144 |
model
|
enriquesaou/phi2_mrqa_v2
| 1 |
[
"main"
] | 104,894,177 |
model
|
med-alex/uzn-roberta-base-ft-qa-tr-mt-to-uzn
| 1 |
[
"main"
] | 352,186,225 |
model
|
paulo037/stable-code-instruct-3b-spider-13000-steps
| 1 |
[
"main"
] | 11,183,918,197 |
model
|
austindavis/gpt2-lichess-uci-201302b
| 1 |
[
"main"
] | 102,099,077 |
model
|
NBA55/Final_DPO_All_3_issue_with_Cleaned_Dataset_epoch_2
| 1 |
[
"main"
] | 201,360,658 |
model
|
RichardErkhov/42dot_-_42dot_LLM-SFT-1.3B-4bits
| 1 |
[
"main"
] | 1,109,709,848 |
model
|
Meggido/Pantheon-RP-1.0-8b-Llama-3-6.5bpw-h8-exl2
| 1 |
[
"main"
] | 7,251,514,066 |
model
|
CyberHarem/sruler_fgo
| 1 |
[
"main"
] | 2,277,439,745 |
model
|
JoshuaAAX/bert-finetuned-ner
| 1 |
[
"main"
] | 431,897,920 |
model
|
RichardErkhov/42dot_-_42dot_LLM-SFT-1.3B-8bits
| 1 |
[
"main"
] | 1,651,222,088 |
model
|
Zannatul/google-vit-base-patch16-384-in21k-batch_16_epoch_4_classes_24_final_withAug_12th_May
| 1 |
[
"main"
] | 344,687,276 |
model
|
abc88767/22c5
| 1 |
[
"main"
] | 3,295,853,121 |
model
|
stafdif/Milf
| 1 |
[
"main"
] | 19,027,847 |
model
|
techie4life/llama-2-7b-boolq
| 1 |
[
"main"
] | 26,956,116,765 |
model
|
soba1911/xlrs-wav2vec2-lr1e-4
| 1 |
[
"main"
] | 3,764,638,114 |
model
|
Ransss/kukuspice-7B-Q8_0-GGUF
| 1 |
[
"main"
] | 7,695,860,369 |
model
|
mberghouse/llama3-8b-python-codes-25k
| 1 |
[
"main"
] | 16,069,719,273 |
model
|
BlackRose104/EpiCRealism
| 1 |
[
"main"
] | 2,134,468,753 |
model
|
RichardErkhov/NousResearch_-_Hermes-2-Pro-Llama-3-8B-gguf
| 1 |
[
"main"
] | 99,039,359,917 |
model
|
sajithrpc/gemma-2b-mt-German-to-English
| 1 |
[
"main"
] | 5,034,185,800 |
model
|
akbargherbal/ar_poem_v_01_lora
| 1 |
[
"main"
] | 221,872,167 |
model
|
abc88767/3sc4
| 1 |
[
"main"
] | 3,295,853,121 |
model
|
5w4n/poneyate-xl-v1
| 1 |
[
"main"
] | 13,878,866,235 |
model
|
alexandra-udrescu/qa_toolformer_2
| 1 |
[
"main"
] | 67,133,370 |
model
|
RichardErkhov/BAAI_-_Bunny-v1_0-3B-4bits
| 1 |
[
"main"
] | 1,941,407,307 |
model
|
abc88767/5c4
| 1 |
[
"main"
] | 3,295,853,120 |
model
|
cpellens/wavecoder-ultra-6.7b-Q4_K_M-GGUF
| 1 |
[
"main"
] | 4,081,565,367 |
model
|
RichardErkhov/BAAI_-_Bunny-v1_0-3B-8bits
| 1 |
[
"main"
] | 3,049,487,475 |
model
|
eitan3/infini_llama_3_v2
| 1 |
[
"main"
] | 16,069,733,072 |
model
|
cisnes68/Carl
| 1 |
[
"main"
] | 1,519 |
model
|
nnheui/gpt2-sft-index-op-500
| 1 |
[
"main"
] | 253,206,785 |
model
|
Grayx/sad_papa_7
| 1 |
[
"main"
] | 16,069,724,863 |
model
|
rahulmak92/mistral-finetuned-alpaca
| 1 |
[
"main"
] | 1,519 |
model
|
parthkher123/Gujarati_Llama_3_Model
| 1 |
[
"main"
] | 176,971,858 |
model
|
AI-001/Jellyfishd
| 1 |
[
"main"
] | 11,098,027 |
model
|
Bertinho24/Sumin
| 1 |
[
"main"
] | 77,519,514 |
model
|
ivykopal/wikiann_sk_adapter_100k
| 1 |
[
"main"
] | 775,538,706 |
model
|
Bertinho24/Sieun
| 1 |
[
"main"
] | 70,840,174 |
model
|
Zannatul/google-vit-base-patch32-224-in21k-batch_16_epoch_4_classes_24
| 1 |
[
"main"
] | 1,519 |
model
|
Bertinho24/Isa
| 1 |
[
"main"
] | 80,404,044 |
model
|
Bertinho24/Seeun
| 1 |
[
"main"
] | 74,751,625 |
model
|
cpellens/granite-8b-code-instruct-Q4_K_M-GGUF
| 1 |
[
"main"
] | 4,882,861,592 |
model
|
Bertinho24/Yoon
| 1 |
[
"main"
] | 78,775,739 |
model
|
santoshtyss/lex-32k-100
| 1 |
[
"main"
] | 28,444,991,914 |
model
|
rohan-2810/image-captioning-blip
| 1 |
[
"main"
] | 5,638,817,966 |
model
|
gauravsirola/my_awesome_eli5_clm-model
| 1 |
[
"main"
] | 327,665,516 |
model
|
Kikiraw/MOONFILM
| 1 |
[
"main"
] | 11,076,427,155 |
model
|
llmsec/Llama-3-8B-sft-lora-ultrachat
| 1 |
[
"main"
] | 227,284,323 |
model
|
Bertinho24/J
| 1 |
[
"main"
] | 80,149,181 |
model
|
oliverkm/translation
| 1 |
[
"main"
] | 1,563 |
model
|
rohan-2810/imagecap_blipbase
| 1 |
[
"main"
] | 1,981,045,889 |
model
|
PlayerBPlaytime/MJ
| 1 |
[
"main"
] | 1,295,311,081 |
model
|
nbeerbower/llama-3-stinky-v2-8B
| 1 |
[
"main"
] | 16,069,731,749 |
model
|
tomaszki/stablelm-57
| 1 |
[
"main"
] | 3,295,880,749 |
model
|
bogya/segformer-b0-scene-parse-150
| 1 |
[
"main"
] | 1,519 |
model
|
omar-sala7/falcon7binstruct_qac_updated_context
| 1 |
[
"main"
] | 263,894,631 |
model
|
meherajj/Bangla-Llama-2-7b-base-v0.1
| 1 |
[
"main"
] | 13,479,236,859 |
model
|
Fralet/personality
| 1 |
[
"main"
] | 6,932,305,260 |
model
|
Miguelsbdh/llama-3-finetuned-bases-de-datos
| 1 |
[
"main"
] | 11,477,845,288 |
model
|
AlignmentResearch/robust_llm_pythia-160m_niki-046_enronspam_random-token-1280_seed-2
| 31 |
[
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model
|
NBA55/Final_Experiment_with_trained_model_Final_DPO_for_all_3_epoch_2_with_cleaned_dataset
| 1 |
[
"main"
] | 25,198,968 |
model
|
Norphel/dzoQA_f2f
| 1 |
[
"main"
] | 431,832,088 |
model
|
ernestsborro/erfdse
| 1 |
[
"main"
] | 1,552 |
model
|
solidrust/Llama-3-monika-ddlc-8b-v1-AWQ
| 1 |
[
"main"
] | 5,737,143,613 |
model
|
mradermacher/Exodius-70B-i1-GGUF
| 1 |
[
"main"
] | 704,070,265,469 |
model
|
nev/phi-3-4k-saex-test
| 1 |
[
"main"
] | 137,798,166,759 |
model
|
lalreddy/HW_7
| 1 |
[
"main"
] | 1,519 |
model
|
iujinasena/llama-2-7b-conv
| 1 |
[
"main"
] | 13,476,896,283 |
model
|
AlignmentResearch/robust_llm_pythia-160m_niki-046_enronspam_random-token-1280_seed-1
| 31 |
[
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model
|
thrunlab/cats_exp
| 1 |
[
"main"
] | 28,967,008,061 |
model
|
xizak/bert-base-uncased-MR
| 1 |
[
"main"
] | 1,543 |
model
|
axiomailabs/lora_model_fine_tuned_llama3_8b
| 1 |
[
"main"
] | 167,835,069 |
model
|
axiomailabs/lora_model_fine_tuned_llama3_8b_tokenizer
| 1 |
[
"main"
] | 9,143,476 |
model
|
MaestraAgah/whisper-large-v3-en
| 1 |
[
"main"
] | 1,519 |
model
|
sahlebrahim/codeparrot-ds
| 1 |
[
"main"
] | 500,348,413 |
model
|
Felladrin/gguf-sharded-vicuna-160m
| 1 |
[
"main"
] | 498,963,465 |
model
|
memoire/llama3_cicis2017_01
| 1 |
[
"main"
] | 4,920,735,799 |
model
|
DewiBrynJones/wav2vec2-xls-r-2b-ft-btb-cy
| 1 |
[
"main"
] | 8,637,694,226 |
model
|
zainalikhokhar/llama2_VLLM_LORA
| 1 |
[
"main"
] | 162,319,417 |
model
|
fundahe/fundi
| 1 |
[
"main"
] | 1,519 |
model
|
zainalikhokhar/llama2_VLLM_4bit
| 1 |
[
"main"
] | 3,868,394,285 |
model
|
zainalikhokhar/llama2_VLLM_16bit
| 1 |
[
"main"
] | 26,431,251,386 |
model
|
RichardErkhov/SanjiWatsuki_-_Kunoichi-DPO-v2-7B-gguf
| 1 |
[
"main"
] | 87,511,270,912 |
model
|
DUAL-GPO/phi-2-gpo-v2-i1
| 1 |
[
"main"
] | 171,307,061 |
model
|
mnoukhov/EleutherAI_pythia-1b-deduped__sft__tldr_55513
| 1 |
[
"main"
] | 4,049,276,940 |
model
|
RichardErkhov/PygmalionAI_-_pygmalion-2-13b-4bits
| 1 |
[
"main"
] | 7,795,699,829 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.