type
stringclasses 1
value | id
stringlengths 5
122
| num_branches
int64 1
1.76k
| branches
listlengths 1
1.76k
| main_branch_size
int64 0
32,943B
|
---|---|---|---|---|
model
|
EGORsGOR/vit-spam
| 1 |
[
"main"
] | 343,242,409 |
model
|
kaya-kedi/Bubble-TADC
| 1 |
[
"main"
] | 54,960,071 |
model
|
Minbyul/llama2-7b-wo-kqa_golden-iter-dpo-step2
| 1 |
[
"main"
] | 13,479,247,039 |
model
|
ccihvun/zziclo
| 1 |
[
"main"
] | 1,519 |
model
|
blockblockblock/Dark-Miqu-70B-bpw6-exl2
| 1 |
[
"main"
] | 52,013,010,661 |
model
|
jidlso/RVC-models
| 1 |
[
"main"
] | 215,983,495 |
model
|
copper029/videomae-base-finetuned-ucf101-subset
| 1 |
[
"main"
] | 1,519 |
model
|
AlignmentResearch/robust_llm_pythia-160m_niki-046_enronspam_random-token-1280_seed-0
| 31 |
[
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model
|
chujiezheng/tulu-2-dpo-13b
| 1 |
[
"main"
] | 26,034,153,661 |
model
|
rhye/man
| 1 |
[
"main"
] | 5,482,649,875 |
model
|
wikiray/llama
| 1 |
[
"main"
] | 1,546 |
model
|
theskrtnerd/corgi-dog
| 1 |
[
"main"
] | 5,482,649,438 |
model
|
AlignmentResearch/robust_llm_pythia-70m_niki-047_wl_random-token-1280_seed-1
| 31 |
[
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model
|
AlignmentResearch/robust_llm_pythia-70m_niki-047_wl_random-token-1280_seed-0
| 31 |
[
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model
|
nickrwu/deberta-v3-large-finetuned-mathqa
| 1 |
[
"main"
] | 1,519 |
model
|
ke-lly/46401345_1
| 1 |
[
"main"
] | 1,519 |
model
|
CarlRD/bloom-7b1-lora-tagger
| 1 |
[
"main"
] | 1,519 |
model
|
DUAL-GPO/zephyr-7b-gpo-v4-i2
| 1 |
[
"main"
] | 673,548,959 |
model
|
Thangaraj/zephyr-support-chatbot
| 1 |
[
"main"
] | 29,585,688 |
model
|
ABF54/testingdulu
| 1 |
[
"main"
] | 1,548 |
model
|
tayyibsupercool/llama-2-7b-miniguanaco-qlora
| 1 |
[
"main"
] | 16,826,198 |
model
|
AdityaXPV/Llama-3-8b-Instruct-law-sage-v0.1-GGUF
| 1 |
[
"main"
] | 37,653,351,430 |
model
|
Mohyuddin1973/Puppy
| 1 |
[
"main"
] | 1,561 |
model
|
Caiyun-AI/DCPythia-6.9B
| 1 |
[
"main"
] | 14,809,074,522 |
model
|
SubhasishSaha/Resume-Classifier
| 1 |
[
"main"
] | 438,995,324 |
model
|
RichardErkhov/YeungNLP_-_firefly-mixtral-8x7b-gguf
| 1 |
[
"main"
] | 246,760,208,286 |
model
|
snshrivas10/zephyr-support-chatbot
| 1 |
[
"main"
] | 29,585,688 |
model
|
bhassi01/zephyr-support-chatbot
| 1 |
[
"main"
] | 29,585,688 |
model
|
sunjupskilling/zephyr-support-chatbot
| 1 |
[
"main"
] | 29,585,943 |
model
|
praveenku32k/zephyr-LLM-chatbot-General
| 1 |
[
"main"
] | 29,585,772 |
model
|
Yosingh1/zephyr-support-chatbot
| 1 |
[
"main"
] | 29,585,943 |
model
|
SachinGenAIMaster/zephyr-support-chatbot
| 1 |
[
"main"
] | 29,585,943 |
model
|
PraveenCMR/zephyr-support-chatbot
| 1 |
[
"main"
] | 29,585,943 |
model
|
ittailup/whisper-small-hi
| 1 |
[
"main"
] | 1,519 |
model
|
Tristan/pythia-70m_default
| 1 |
[
"main"
] | 283,812,633 |
model
|
ydavidpan/bert-base-uncased-finetuned-kaggle-arena-0510
| 1 |
[
"main"
] | 1,519 |
model
|
SaravanaPriyan/zephyr-support-chatbot
| 1 |
[
"main"
] | 29,585,688 |
model
|
ivilson/llama-3-8b-chinese-function-calling-AWQ-4bit
| 1 |
[
"main"
] | 5,737,141,885 |
model
|
med-alex/kaz-roberta-base-ft-qa-en-mt-to-kaz
| 1 |
[
"main"
] | 368,017,203 |
model
|
theskrtnerd/cocacola-drink
| 1 |
[
"main"
] | 5,482,649,460 |
model
|
Ayush-1722/Llama-3-8B-Instruct-v0.1-32K-QLoRA-Merged
| 1 |
[
"main"
] | 16,069,789,399 |
model
|
Kearm/file.pt
| 1 |
[
"main"
] | 1,519 |
model
|
smorodinovmv/models1
| 1 |
[
"main"
] | 1,519 |
model
|
aariz120/zephyr-support-chatbot
| 1 |
[
"main"
] | 29,592,464 |
model
|
arctic126/lab_h2o_35
| 1 |
[
"main"
] | 3,664,812,902 |
model
|
leimu/44
| 1 |
[
"main"
] | 228,455,987 |
model
|
npedrazzini/BERTislav
| 1 |
[
"main"
] | 1,432,810,330 |
model
|
jujusosmart/bert-base-chinese-Guilt
| 1 |
[
"main"
] | 409,107,863 |
model
|
Luna-Skywalker/phi-3-finetuned-lora
| 1 |
[
"main"
] | 120,104,109 |
model
|
blockblockblock/neo_7b-bpw2.25-exl2
| 1 |
[
"main"
] | 2,634,083,599 |
model
|
desi3/ppo-LunarLander-v2
| 1 |
[
"main"
] | 428,244 |
model
|
omark03/llama2-qlora-finetunined-med-convo-v1
| 1 |
[
"main"
] | 1,519 |
model
|
vinitlondhe21/zephyr-support-chatbot
| 1 |
[
"main"
] | 29,585,688 |
model
|
leimu/45
| 1 |
[
"main"
] | 228,458,107 |
model
|
Jackie999/mistral-10p-adv
| 1 |
[
"main"
] | 32,101,344 |
model
|
Aspik101/llama1111
| 1 |
[
"main"
] | 32,130,281,223 |
model
|
AlignmentResearch/robust_llm_pythia-410m_niki-047_wl_random-token-1280_seed-2
| 31 |
[
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model
|
SimplCup/OlivisRodrigo
| 1 |
[
"main"
] | 215,944,159 |
model
|
AlignmentResearch/robust_llm_pythia-70m_niki-047_wl_random-token-1280_seed-2
| 31 |
[
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model
|
AlignmentResearch/robust_llm_pythia-160m_niki-047_wl_random-token-1280_seed-0
| 31 |
[
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model
|
rhye/cat_ppl_less
| 1 |
[
"main"
] | 15,797,585,024 |
model
|
mmnga/Meta-Llama-3-8B-Instruct-gguf
| 1 |
[
"main"
] | 99,819,606,841 |
model
|
sachi020991/my_awesome_model
| 1 |
[
"main"
] | 268,801,869 |
model
|
LordNoah/schedule-iter0-b
| 1 |
[
"main"
] | 28,968,791,095 |
model
|
thorirhrafn/GPT1B_domar_RLHF_e3
| 1 |
[
"main"
] | 50,345,320 |
model
|
fine-tuned/jina-embeddings-v2-base-en-5122024-3toh-webapp
| 1 |
[
"main"
] | 550,554,550 |
model
|
Mr-Bhaskar/fbt-gemma-7b-inst
| 1 |
[
"main"
] | 17,093,038,793 |
model
|
acmench/testcopy
| 1 |
[
"main"
] | 311,838,132 |
model
|
blockblockblock/neo_7b-bpw2.5-exl2
| 1 |
[
"main"
] | 2,862,572,846 |
model
|
MrOvkill/Llama-3-8B-sft-lora-ultrachat
| 1 |
[
"main"
] | 1,519 |
model
|
AlignmentResearch/robust_llm_pythia-410m_niki-047_wl_random-token-1280_seed-0
| 31 |
[
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model
|
sally9805/bert-base-uncased-finetuned-news-1908-1909
| 1 |
[
"main"
] | 439,048,284 |
model
|
AlignmentResearch/robust_llm_pythia-1b_niki-047_wl_random-token-1280_seed-1
| 31 |
[
"adv-training-round-29",
"adv-training-round-28",
"adv-training-round-27",
"adv-training-round-26",
"adv-training-round-25",
"adv-training-round-24",
"adv-training-round-23",
"adv-training-round-22",
"adv-training-round-21",
"adv-training-round-20",
"adv-training-round-19",
"adv-training-round-18",
"adv-training-round-17",
"adv-training-round-16",
"adv-training-round-15",
"adv-training-round-14",
"adv-training-round-13",
"adv-training-round-12",
"adv-training-round-11",
"adv-training-round-10",
"adv-training-round-9",
"adv-training-round-8",
"adv-training-round-7",
"adv-training-round-6",
"adv-training-round-5",
"adv-training-round-4",
"adv-training-round-3",
"adv-training-round-2",
"adv-training-round-1",
"adv-training-round-0",
"main"
] | 1,519 |
model
|
techsivam/zephyr-support-chatbot
| 1 |
[
"main"
] | 1,519 |
model
|
yaojingguo/codeparrot-ds
| 1 |
[
"main"
] | 1,519 |
model
|
med-alex/kaz-roberta-base-ft-qa-tr-mt-to-kaz
| 1 |
[
"main"
] | 368,348,009 |
model
|
krishnakalyan3/emo_nmlaozqe_67
| 1 |
[
"main"
] | 382,272,147 |
model
|
DerekTrayn15/theolasad
| 1 |
[
"main"
] | 97,432,764 |
model
|
Rohandebnath243536287/Llama-2-7b-chat-finetune
| 1 |
[
"main"
] | 13,479,328,407 |
model
|
Demon1221/Temp
| 1 |
[
"main"
] | 1,547 |
model
|
win10/phi3-128k-6b
| 1 |
[
"main"
] | 11,721,743,675 |
model
|
voxmenthe/Meta-Llama-3-70B-Instruct-8bit
| 1 |
[
"main"
] | 159,603,261,670 |
model
|
clxudiajazmin/summarization-TFM-Claudia
| 1 |
[
"main"
] | 245,315,332 |
model
|
PB7-DUT-2023/finetuned_Bloomz_1b1_v6
| 1 |
[
"main"
] | 2,145,231,559 |
model
|
nguyenx/vietnamese-correction-v2
| 1 |
[
"main"
] | 1,588,918,422 |
model
|
TrgTuan10/Famelink_hand_negative
| 1 |
[
"main"
] | 28,077 |
model
|
jiangcongtao/llama3-8b-unsloth-merged
| 1 |
[
"main"
] | 16,069,789,678 |
model
|
SuiGio/roberta_pubmesh
| 1 |
[
"main"
] | 331,930,175 |
model
|
blockblockblock/neo_7b-bpw3-exl2
| 1 |
[
"main"
] | 3,323,677,998 |
model
|
Mag0g/Ezekiel25_15
| 1 |
[
"main"
] | 3,295,853,092 |
model
|
Hyx66/LandscaperPriv
| 1 |
[
"main"
] | 2,607,440,397 |
model
|
cwei13/bert-base-japanese-ghost_rate
| 1 |
[
"main"
] | 442,892,896 |
model
|
OwOpeepeepoopoo/test12
| 1 |
[
"main"
] | 3,295,853,400 |
model
|
S4nto/lora-dpo-finetuned-stage2-iter40000-sft
| 1 |
[
"main"
] | 23,066,479,892 |
model
|
jiangcongtao/llama3-8b-unsloth
| 1 |
[
"main"
] | 167,835,070 |
model
|
enchan1/q-FrozenLake-v1-4x4-noSlippery
| 1 |
[
"main"
] | 34,548 |
model
|
Mudiit4/gemma-Code-aws-Finetune-test
| 1 |
[
"main"
] | 5,034,185,180 |
model
|
jin-cheon/bert-finetuned-ner
| 1 |
[
"main"
] | 862,797,421 |
model
|
Mr-Bhaskar/fbt-mistral7b-instruct
| 1 |
[
"main"
] | 14,485,390,390 |
model
|
netcat420/MFANNv0.9
| 1 |
[
"main"
] | 32,130,241,100 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.