text
stringlengths 15
129
|
---|
model:mlx-community/magnum-v2-72b [33] |
model:DouglasBraga/swin-tiny-patch4-window7-224-finetuned-leukemia.v2.0 [10] |
model:sert121/orpo_pythia2_hh_filtered [33] |
model:Paranchai/wav2vec2-large-xlsr-53-th-speech-emotion-recognition-3c-10ep [1] |
model:PygTesting/pyg3v1-nemo-3ep-ckpts [33] |
model:a2ronprice/aaron-lora-2000 [37] |
model:timlenardo/timl_varied_10_realistic_vision_v6.0_B1_noVAE_dreambooth_2500_steps_ohwx [37] |
model:chgk13/Qwen2-0.5B-openvino-8bit [33] |
model:omareldessouky/autotrain-clro7-jl3ak [33] |
model:chgk13/Qwen2-0.5B-openvino [33] |
model:ychu612/BiomedBERT_rsavav_fn_adult2_hq [32] |
model:bethanywork/joshbad [37] |
model:TiziIsa93/ai-velourstrainer [37] |
model:ApprikatAI/Qwen2.5-Coder-1.5B-Instruct-FP16-GGUF [33] |
model:peppescollo/drm_leo_glitter_v2 [37] |
model:sert121/orpo_opt2_hh_filtered [33] |
model:Youlln/ECE-EIFFEL.ia-0.5B-FT-V2-Q8 [33] |
model:timlenardo/timl_varied_10_realistic_vision_v6.0_B1_noVAE_dreambooth_500_steps_ohwx [37] |
model:sheldonrobinson/Llama-3.2-3B-Instruct-Q4_0-GGUF [33] |
model:sheldonrobinson/Qwen2.5-Math-7B-Instruct-Q4_K_M-GGUF [33] |
model:sheldonrobinson/Qwen2.5-Math-1.5B-Instruct-Q4_K_M-GGUF [33] |
model:Iberianson/4b4l02 [37] |
model:dane12341/Be-a-gov [37] |
model:sheldonrobinson/Llama-Guard-3-1B-Q4_0-GGUF [33] |
model:erich01/lacaBottle [37] |
model:sheldonrobinson/all-MiniLM-L12-v2-Q8_0-GGUF [25, 7] |
model:Youlln/ECE.EIFFEIL.ia-0.5B-SLERP [33] |
model:rcastrovexler/whisper-small-es-cl [4] |
model:yscode/calm3-22b-chat-neuronx [33] |
model:timlenardo/timl_varied_10_realistic_vision_v6.0_B1_noVAE_dreambooth_1500_steps_validation_ohwx [37] |
model:bhuv1-c/db-valid-warehouse-sdxl [37] |
model:nanda101010/primeiroprojeto [22] |
model:ekrombouts/zuster_fietje [33] |
model:davidrd123/Flux-MoonLanding-Man-AiToolkit [37] |
model:bbennett17/Llama-3.1-8B-Instruct-Q5_K_M-GGUF [33] |
model:its-magick/merlin-tool-time [37] |
model:nell123/phi-avg-v02 [33] |
model:Moustapha91/speecht5_tts_wolof [36] |
model:cheloricardo/chelopics [37] |
model:gpbhupinder/glitter [37] |
model:gpbhupinder/glitter1 [37] |
model:timlenardo/timl_varied_10_realistic_vision_v2.0_dreambooth_lora_500_steps_validation_ohwx [37] |
model:erich01/lacaBottle2 [37] |
dataset:HuggingFaceFW/fineweb [33] |
dataset:AI-MO/NuminaMath-CoT [33] |
dataset:nkp37/OpenVid-1M [39] |
dataset:HuggingFaceFW/fineweb-edu [33] |
dataset:TIGER-Lab/MMLU-Pro [22] |
dataset:simon3000/genshin-voice [1, 4, 38] |
dataset:allenai/WildChat-1M [33, 22, 40] |
dataset:UCSC-VLAA/MedTrinity-25M [22] |
dataset:BAAI/Infinity-Instruct [33] |
dataset:HuggingFaceM4/Docmatix [47] |
dataset:allenai/wildguardmix [32] |
dataset:proj-persona/PersonaHub [33, 32, 42, 8, 27, 40] |
dataset:mlabonne/orpo-dpo-mix-40k [33] |
dataset:tanganke/stanford_cars [10] |
dataset:BUAADreamer/llava-en-zh-300k [33, 47] |
dataset:TIGER-Lab/WebInstructSub [22] |
dataset:ShareGPT4Video/ShareGPT4Video [47, 22] |
dataset:OpenGVLab/ShareGPT-4o [47, 22] |
dataset:BestWishYsh/ChronoMagic-ProH [39] |
dataset:knowledgator/GLINER-multi-task-synthetic-data [42] |
dataset:NousResearch/hermes-function-calling-v1 [33, 22, 7] |
dataset:MohamedRashad/midjourney-detailed-prompts [37] |
dataset:PleIAs/YouTube-Commons [33] |
dataset:Iker/Document-Translation-en-es [43] |
dataset:Iker/OpenHermes-2.5-Spanish [33] |
dataset:MaziyarPanahi/WizardLM_evol_instruct_V2_196k [33, 22] |
dataset:rootsautomation/RICO-ScreenQA [22] |
dataset:nvidia/Aegis-AI-Content-Safety-Dataset-1.0 [32] |
dataset:argilla/Capybara-Preferences [33] |
dataset:HPAI-BSC/CareQA [22] |
dataset:Shekswess/medical_llama3_instruct_dataset_short [22] |
dataset:TFMC/imatrix-dataset-for-japanese-llm [33] |
dataset:vutuka/aya_african_alpaca [33] |
dataset:common-canvas/commoncatalog-cc-by [37] |
dataset:mhan/Shot2Story-20K [26, 47, 22] |
dataset:reciperesearch/dolphin-sft-v0.1-preference [33] |
dataset:mhan/shot2story [26, 47, 22] |
dataset:sentence-transformers/all-nli [7, 25] |
dataset:sentence-transformers/stsb [7, 25] |
dataset:TIGER-Lab/VideoFeedback [45] |
dataset:masakhane/afrimmlu [22] |
dataset:jojo0217/korean_safe_conversation [33] |
dataset:nthakur/swim-ir-monolingual [34, 22] |
dataset:flozi00/asr-german-mixed [4] |
dataset:Abhaykoul/Ancient-Indian-Wisdom [33] |
dataset:xiaodongguaAIGC/alpaca_en_zh_ruozhiba [33, 40] |
dataset:MetaAligner/HH-RLHF-MetaAligner-Data [33] |
dataset:sentence-transformers/parallel-sentences-talks [7, 25] |
dataset:google/imageinwords [16, 37, 21] |
dataset:sentence-transformers/sentence-compression [7, 25] |
dataset:taddeusb90/finbro-v0.1.0 [32, 22, 26] |
dataset:barsbold/ImagetotextMGL [16] |
dataset:HachiML/Hachi-Alpaca [33] |
dataset:Kongongong/Thai-Physics-Data-40K [22] |
dataset:sentence-transformers/msmarco-msmarco-distilbert-base-tas-b [7, 25] |
dataset:sentence-transformers/msmarco-msmarco-distilbert-base-v3 [7, 25] |
dataset:occiglot/occiglot-fineweb-v0.5 [33] |
Subsets and Splits