modelId
stringlengths
5
139
author
stringlengths
2
42
last_modified
timestamp[us, tz=UTC]date
2020-02-15 11:33:14
2025-09-21 12:34:09
downloads
int64
0
223M
likes
int64
0
11.7k
library_name
stringclasses
568 values
tags
listlengths
1
4.05k
pipeline_tag
stringclasses
55 values
createdAt
timestamp[us, tz=UTC]date
2022-03-02 23:29:04
2025-09-21 12:33:58
card
stringlengths
11
1.01M
ultratopaz/85554
ultratopaz
2025-08-19T22:09:31Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:09:29Z
[View on Civ Archive](https://civarchive.com/models/110731?modelVersionId=119395)
ultratopaz/38766
ultratopaz
2025-08-19T22:09:24Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:09:21Z
[View on Civ Archive](https://civarchive.com/models/48793?modelVersionId=53376)
ultratopaz/100911
ultratopaz
2025-08-19T22:09:17Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:09:14Z
[View on Civ Archive](https://civarchive.com/models/126037?modelVersionId=137746)
crystalline7/39528
crystalline7
2025-08-19T22:08:55Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:08:52Z
[View on Civ Archive](https://civarchive.com/models/50237?modelVersionId=54765)
seraphimzzzz/422167
seraphimzzzz
2025-08-19T22:08:36Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:08:30Z
[View on Civ Archive](https://civarchive.com/models/393921?modelVersionId=504614)
Muapi/illulora-simple-illustration-style
Muapi
2025-08-19T22:08:10Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-19T22:07:57Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # IlluLoRA - Simple Illustration Style ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: IlluLORA ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:665753@745084", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
seraphimzzzz/98448
seraphimzzzz
2025-08-19T22:08:03Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:07:33Z
[View on Civ Archive](https://civarchive.com/models/123650?modelVersionId=134834)
Kurosawama/Llama-3.2-3B-Full-align
Kurosawama
2025-08-19T22:07:56Z
0
0
transformers
[ "transformers", "safetensors", "trl", "dpo", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2025-08-19T22:07:49Z
--- library_name: transformers tags: - trl - dpo --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
Hobaks/Qwen3-30B-A3B-Instruct-2507-Q4_K_M-GGUF
Hobaks
2025-08-19T22:07:51Z
0
0
transformers
[ "transformers", "gguf", "llama-cpp", "gguf-my-repo", "text-generation", "base_model:Qwen/Qwen3-30B-A3B-Instruct-2507", "base_model:quantized:Qwen/Qwen3-30B-A3B-Instruct-2507", "license:apache-2.0", "endpoints_compatible", "region:us", "conversational" ]
text-generation
2025-08-19T22:06:34Z
--- library_name: transformers license: apache-2.0 license_link: https://huggingface.co/Qwen/Qwen3-30B-A3B-Instruct-2507/blob/main/LICENSE pipeline_tag: text-generation base_model: Qwen/Qwen3-30B-A3B-Instruct-2507 tags: - llama-cpp - gguf-my-repo --- # Hobaks/Qwen3-30B-A3B-Instruct-2507-Q4_K_M-GGUF This model was converted to GGUF format from [`Qwen/Qwen3-30B-A3B-Instruct-2507`](https://huggingface.co/Qwen/Qwen3-30B-A3B-Instruct-2507) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space. Refer to the [original model card](https://huggingface.co/Qwen/Qwen3-30B-A3B-Instruct-2507) for more details on the model. ## Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) ```bash brew install llama.cpp ``` Invoke the llama.cpp server or the CLI. ### CLI: ```bash llama-cli --hf-repo Hobaks/Qwen3-30B-A3B-Instruct-2507-Q4_K_M-GGUF --hf-file qwen3-30b-a3b-instruct-2507-q4_k_m.gguf -p "The meaning to life and the universe is" ``` ### Server: ```bash llama-server --hf-repo Hobaks/Qwen3-30B-A3B-Instruct-2507-Q4_K_M-GGUF --hf-file qwen3-30b-a3b-instruct-2507-q4_k_m.gguf -c 2048 ``` Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well. Step 1: Clone llama.cpp from GitHub. ``` git clone https://github.com/ggerganov/llama.cpp ``` Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux). ``` cd llama.cpp && LLAMA_CURL=1 make ``` Step 3: Run inference through the main binary. ``` ./llama-cli --hf-repo Hobaks/Qwen3-30B-A3B-Instruct-2507-Q4_K_M-GGUF --hf-file qwen3-30b-a3b-instruct-2507-q4_k_m.gguf -p "The meaning to life and the universe is" ``` or ``` ./llama-server --hf-repo Hobaks/Qwen3-30B-A3B-Instruct-2507-Q4_K_M-GGUF --hf-file qwen3-30b-a3b-instruct-2507-q4_k_m.gguf -c 2048 ```
seraphimzzzz/44530
seraphimzzzz
2025-08-19T22:07:18Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:07:14Z
[View on Civ Archive](https://civarchive.com/models/58763?modelVersionId=63204)
crystalline7/55160
crystalline7
2025-08-19T22:06:47Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:06:47Z
[View on Civ Archive](https://civarchive.com/models/75657?modelVersionId=80415)
Muapi/fictional-model-sabrina-flux
Muapi
2025-08-19T22:06:44Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-19T22:06:30Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # Fictional Model Sabrina - FLUX ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: FictionalSabrina, red head girl with green eyes ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:1088990@1223005", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
seraphimzzzz/481012
seraphimzzzz
2025-08-19T22:06:41Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:06:35Z
[View on Civ Archive](https://civarchive.com/models/498376?modelVersionId=554000)
crystalline7/55386
crystalline7
2025-08-19T22:06:20Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:06:16Z
[View on Civ Archive](https://civarchive.com/models/75729?modelVersionId=80767)
crystalline7/79791
crystalline7
2025-08-19T22:05:59Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:05:55Z
[View on Civ Archive](https://civarchive.com/models/18663?modelVersionId=112521)
seraphimzzzz/93417
seraphimzzzz
2025-08-19T22:05:00Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:04:44Z
[View on Civ Archive](https://civarchive.com/models/118808?modelVersionId=128939)
chansung/Gemma2-2B-CCRL-CUR-COMPLEX-ONLY-1E
chansung
2025-08-19T22:04:44Z
0
0
transformers
[ "transformers", "safetensors", "gemma2", "text-generation", "generated_from_trainer", "open-r1", "trl", "grpo", "conversational", "dataset:chansung/verifiable-coding-problems-python-v2", "arxiv:2402.03300", "base_model:google/gemma-2-2b-it", "base_model:finetune:google/gemma-2-2b-it", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-08-19T02:44:25Z
--- base_model: google/gemma-2-2b-it datasets: chansung/verifiable-coding-problems-python-v2 library_name: transformers model_name: Gemma2-2B-CCRL-CUR-COMPLEX-ONLY-1E tags: - generated_from_trainer - open-r1 - trl - grpo licence: license --- # Model Card for Gemma2-2B-CCRL-CUR-COMPLEX-ONLY-1E This model is a fine-tuned version of [google/gemma-2-2b-it](https://huggingface.co/google/gemma-2-2b-it) on the [chansung/verifiable-coding-problems-python-v2](https://huggingface.co/datasets/chansung/verifiable-coding-problems-python-v2) dataset. It has been trained using [TRL](https://github.com/huggingface/trl). ## Quick start ```python from transformers import pipeline question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?" generator = pipeline("text-generation", model="chansung/Gemma2-2B-CCRL-CUR-COMPLEX-ONLY-1E", device="cuda") output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0] print(output["generated_text"]) ``` ## Training procedure [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/chansung18/huggingface/runs/q0xteiho) This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300). ### Framework versions - TRL: 0.18.0.dev0 - Transformers: 4.52.0.dev0 - Pytorch: 2.6.0 - Datasets: 4.0.0 - Tokenizers: 0.21.4 ## Citations Cite GRPO as: ```bibtex @article{zhihong2024deepseekmath, title = {{DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models}}, author = {Zhihong Shao and Peiyi Wang and Qihao Zhu and Runxin Xu and Junxiao Song and Mingchuan Zhang and Y. K. Li and Y. Wu and Daya Guo}, year = 2024, eprint = {arXiv:2402.03300}, } ``` Cite TRL as: ```bibtex @misc{vonwerra2022trl, title = {{TRL: Transformer Reinforcement Learning}}, author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallou{\'e}dec}, year = 2020, journal = {GitHub repository}, publisher = {GitHub}, howpublished = {\url{https://github.com/huggingface/trl}} } ```
crystalline7/32214
crystalline7
2025-08-19T22:03:58Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:03:55Z
[View on Civ Archive](https://civarchive.com/models/35788?modelVersionId=41989)
Muapi/art-nouveau-flux-lora
Muapi
2025-08-19T22:03:53Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-19T22:03:40Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # Art Nouveau - Flux Lora ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: art nouveau illustration, vintage ( no need specific key word to work ) ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:638308@714072", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
seraphimzzzz/87766
seraphimzzzz
2025-08-19T22:03:51Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:03:49Z
[View on Civ Archive](https://civarchive.com/models/109244?modelVersionId=122008)
seraphimzzzz/15266
seraphimzzzz
2025-08-19T22:03:36Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:03:33Z
[View on Civ Archive](https://civarchive.com/models/14816?modelVersionId=18242)
ultratopaz/20170
ultratopaz
2025-08-19T22:03:27Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:03:23Z
[View on Civ Archive](https://civarchive.com/models/20449?modelVersionId=24314)
seraphimzzzz/559245
seraphimzzzz
2025-08-19T22:03:09Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:03:03Z
[View on Civ Archive](https://civarchive.com/models/577873?modelVersionId=644417)
Muapi/emberveil-semi-realistic-digital-painting
Muapi
2025-08-19T22:02:42Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-19T22:02:18Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # Emberveil | semi-realistic digital painting ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:1050516@1178753", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
seraphimzzzz/805043
seraphimzzzz
2025-08-19T22:02:33Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:02:30Z
[View on Civ Archive](https://civarchive.com/models/799025?modelVersionId=893455)
ultratopaz/9601
ultratopaz
2025-08-19T22:02:25Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:02:22Z
[View on Civ Archive](https://civarchive.com/models/8453?modelVersionId=9969)
mradermacher/QiMing-Holos-Plus-4B-GGUF
mradermacher
2025-08-19T22:02:18Z
0
0
transformers
[ "transformers", "gguf", "qwen", "qwen3", "unsloth", "qiming", "qiming-holos", "bagua", "decision-making", "strategic-analysis", "cognitive-architecture", "chat", "lora", "philosophy-driven-ai", "zh", "en", "base_model:aifeifei798/QiMing-Holos-Plus-4B", "base_model:adapter:aifeifei798/QiMing-Holos-Plus-4B", "license:apache-2.0", "endpoints_compatible", "region:us", "conversational" ]
null
2025-08-19T20:13:11Z
--- base_model: aifeifei798/QiMing-Holos-Plus-4B language: - zh - en library_name: transformers license: apache-2.0 mradermacher: readme_rev: 1 quantized_by: mradermacher tags: - qwen - qwen3 - unsloth - qiming - qiming-holos - bagua - decision-making - strategic-analysis - cognitive-architecture - chat - lora - philosophy-driven-ai --- ## About <!-- ### quantize_version: 2 --> <!-- ### output_tensor_quantised: 1 --> <!-- ### convert_type: hf --> <!-- ### vocab_type: --> <!-- ### tags: --> <!-- ### quants: x-f16 Q4_K_S Q2_K Q8_0 Q6_K Q3_K_M Q3_K_S Q3_K_L Q4_K_M Q5_K_S Q5_K_M IQ4_XS --> <!-- ### quants_skip: --> <!-- ### skip_mmproj: --> static quants of https://huggingface.co/aifeifei798/QiMing-Holos-Plus-4B <!-- provided-files --> ***For a convenient overview and download list, visit our [model page for this model](https://hf.tst.eu/model#QiMing-Holos-Plus-4B-GGUF).*** weighted/imatrix quants are available at https://huggingface.co/mradermacher/QiMing-Holos-Plus-4B-i1-GGUF ## Usage If you are unsure how to use GGUF files, refer to one of [TheBloke's READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for more details, including on how to concatenate multi-part files. ## Provided Quants (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | [GGUF](https://huggingface.co/mradermacher/QiMing-Holos-Plus-4B-GGUF/resolve/main/QiMing-Holos-Plus-4B.Q2_K.gguf) | Q2_K | 1.8 | | | [GGUF](https://huggingface.co/mradermacher/QiMing-Holos-Plus-4B-GGUF/resolve/main/QiMing-Holos-Plus-4B.Q3_K_S.gguf) | Q3_K_S | 2.0 | | | [GGUF](https://huggingface.co/mradermacher/QiMing-Holos-Plus-4B-GGUF/resolve/main/QiMing-Holos-Plus-4B.Q3_K_M.gguf) | Q3_K_M | 2.2 | lower quality | | [GGUF](https://huggingface.co/mradermacher/QiMing-Holos-Plus-4B-GGUF/resolve/main/QiMing-Holos-Plus-4B.Q3_K_L.gguf) | Q3_K_L | 2.3 | | | [GGUF](https://huggingface.co/mradermacher/QiMing-Holos-Plus-4B-GGUF/resolve/main/QiMing-Holos-Plus-4B.IQ4_XS.gguf) | IQ4_XS | 2.4 | | | [GGUF](https://huggingface.co/mradermacher/QiMing-Holos-Plus-4B-GGUF/resolve/main/QiMing-Holos-Plus-4B.Q4_K_S.gguf) | Q4_K_S | 2.5 | fast, recommended | | [GGUF](https://huggingface.co/mradermacher/QiMing-Holos-Plus-4B-GGUF/resolve/main/QiMing-Holos-Plus-4B.Q4_K_M.gguf) | Q4_K_M | 2.6 | fast, recommended | | [GGUF](https://huggingface.co/mradermacher/QiMing-Holos-Plus-4B-GGUF/resolve/main/QiMing-Holos-Plus-4B.Q5_K_S.gguf) | Q5_K_S | 2.9 | | | [GGUF](https://huggingface.co/mradermacher/QiMing-Holos-Plus-4B-GGUF/resolve/main/QiMing-Holos-Plus-4B.Q5_K_M.gguf) | Q5_K_M | 3.0 | | | [GGUF](https://huggingface.co/mradermacher/QiMing-Holos-Plus-4B-GGUF/resolve/main/QiMing-Holos-Plus-4B.Q6_K.gguf) | Q6_K | 3.4 | very good quality | | [GGUF](https://huggingface.co/mradermacher/QiMing-Holos-Plus-4B-GGUF/resolve/main/QiMing-Holos-Plus-4B.Q8_0.gguf) | Q8_0 | 4.4 | fast, best quality | | [GGUF](https://huggingface.co/mradermacher/QiMing-Holos-Plus-4B-GGUF/resolve/main/QiMing-Holos-Plus-4B.f16.gguf) | f16 | 8.2 | 16 bpw, overkill | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): ![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 ## FAQ / Model Request See https://huggingface.co/mradermacher/model_requests for some answers to questions you might have and/or if you want some other model quantized. ## Thanks I thank my company, [nethype GmbH](https://www.nethype.de/), for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. <!-- end -->
crystalline7/892165
crystalline7
2025-08-19T22:01:56Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:01:53Z
[View on Civ Archive](https://civarchive.com/models/879759?modelVersionId=984836)
seraphimzzzz/14697
seraphimzzzz
2025-08-19T22:01:49Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:01:47Z
[View on Civ Archive](https://civarchive.com/models/14867?modelVersionId=17515)
Muapi/cyberpunk-style-enhancer-flux
Muapi
2025-08-19T22:01:46Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-19T22:01:29Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # 🌀 Cyberpunk Style Enhancer [Flux] ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:890818@996849", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
ultratopaz/56525
ultratopaz
2025-08-19T22:01:41Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:01:37Z
[View on Civ Archive](https://civarchive.com/models/44324?modelVersionId=82580)
ultratopaz/48276
ultratopaz
2025-08-19T22:01:16Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:01:11Z
[View on Civ Archive](https://civarchive.com/models/64415?modelVersionId=69038)
Muapi/alien-landscapes
Muapi
2025-08-19T22:01:10Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-19T22:00:54Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # Alien Landscapes ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:562478@745894", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
Leoar/blockassist-bc-pudgy_toothy_cheetah_1755640732
Leoar
2025-08-19T22:01:09Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "pudgy toothy cheetah", "arxiv:2504.07091", "region:us" ]
null
2025-08-19T22:00:58Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - pudgy toothy cheetah --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
crystalline7/91646
crystalline7
2025-08-19T22:00:54Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:00:50Z
[View on Civ Archive](https://civarchive.com/models/73936?modelVersionId=125362)
seraphimzzzz/54659
seraphimzzzz
2025-08-19T22:00:44Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:00:40Z
[View on Civ Archive](https://civarchive.com/models/73936?modelVersionId=79631)
Muapi/xenomorph-xl-sd1.5-f1d
Muapi
2025-08-19T22:00:44Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-19T21:58:51Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # Xenomorph XL + SD1.5 + F1D ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: Xenomorph style ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:388478@1105778", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
seraphimzzzz/99540
seraphimzzzz
2025-08-19T22:00:35Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:00:32Z
[View on Civ Archive](https://civarchive.com/models/124733?modelVersionId=136220)
seraphimzzzz/113286
seraphimzzzz
2025-08-19T22:00:22Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:00:20Z
[View on Civ Archive](https://civarchive.com/models/137778?modelVersionId=152138)
Patzark/wav2vec2-finetuned-portuguese
Patzark
2025-08-19T22:00:17Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "wav2vec2", "automatic-speech-recognition", "generated_from_trainer", "base_model:facebook/wav2vec2-large-xlsr-53", "base_model:finetune:facebook/wav2vec2-large-xlsr-53", "license:apache-2.0", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2025-08-19T05:35:58Z
--- library_name: transformers license: apache-2.0 base_model: facebook/wav2vec2-large-xlsr-53 tags: - generated_from_trainer model-index: - name: wav2vec2-finetuned-portuguese results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # wav2vec2-finetuned-portuguese This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - num_epochs: 10 - mixed_precision_training: Native AMP ### Training results ### Framework versions - Transformers 4.55.2 - Pytorch 2.8.0+cu126 - Datasets 4.0.0 - Tokenizers 0.21.4
ultratopaz/177423
ultratopaz
2025-08-19T22:00:15Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:00:11Z
[View on Civ Archive](https://civarchive.com/models/204283?modelVersionId=230017)
seraphimzzzz/11524
seraphimzzzz
2025-08-19T22:00:05Z
0
0
null
[ "region:us" ]
null
2025-08-19T22:00:01Z
[View on Civ Archive](https://civarchive.com/models/10760?modelVersionId=12772)
crystalline7/49563
crystalline7
2025-08-19T21:59:50Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:59:48Z
[View on Civ Archive](https://civarchive.com/models/66565?modelVersionId=71222)
ultratopaz/67509
ultratopaz
2025-08-19T21:59:38Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:59:35Z
[View on Civ Archive](https://civarchive.com/models/91128?modelVersionId=97121)
seraphimzzzz/57053
seraphimzzzz
2025-08-19T21:59:31Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:59:29Z
[View on Civ Archive](https://civarchive.com/models/78652?modelVersionId=83437)
ultratopaz/66544
ultratopaz
2025-08-19T21:59:10Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:59:08Z
[View on Civ Archive](https://civarchive.com/models/90088?modelVersionId=95932)
ultratopaz/39149
ultratopaz
2025-08-19T21:59:04Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:59:02Z
[View on Civ Archive](https://civarchive.com/models/49489?modelVersionId=54066)
ultratopaz/72344
ultratopaz
2025-08-19T21:58:57Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:58:55Z
[View on Civ Archive](https://civarchive.com/models/48727?modelVersionId=103126)
crystalline7/80244
crystalline7
2025-08-19T21:58:51Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:58:49Z
[View on Civ Archive](https://civarchive.com/models/105393?modelVersionId=113058)
crystalline7/49570
crystalline7
2025-08-19T21:58:45Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:58:43Z
[View on Civ Archive](https://civarchive.com/models/66573?modelVersionId=71230)
ultratopaz/39163
ultratopaz
2025-08-19T21:58:39Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:58:36Z
[View on Civ Archive](https://civarchive.com/models/49522?modelVersionId=54098)
calegpedia/blockassist-bc-stealthy_slimy_rooster_1755639107
calegpedia
2025-08-19T21:58:24Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "stealthy slimy rooster", "arxiv:2504.07091", "region:us" ]
null
2025-08-19T21:58:21Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - stealthy slimy rooster --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
Muapi/araminta-s-glamourphotography-sdxl-flux
Muapi
2025-08-19T21:58:02Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-19T21:57:45Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # araminta-s-glamourphotography (SDXL+Flux) ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:582369@772166", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
ultratopaz/58208
ultratopaz
2025-08-19T21:57:58Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:57:55Z
[View on Civ Archive](https://civarchive.com/models/80397?modelVersionId=85247)
seraphimzzzz/46722
seraphimzzzz
2025-08-19T21:57:30Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:57:30Z
[View on Civ Archive](https://civarchive.com/models/62174?modelVersionId=66712)
Muapi/flux.1-d-realistic-genshin-impact-cosplay-official-doujin-costume-collection-cosplay
Muapi
2025-08-19T21:57:24Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-19T21:57:08Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # [Flux.1 D][Realistic] <Genshin Impact>Cosplay(official/doujin) costume collection|原神cosplay(官设/同人)服装集合 ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: A realistic photo of a tall and slender beautiful young woman in cyb-skirk cosplay costume. She is also wearing tight elbow gloves and tight thighhighs and cosplay high heel boots. She has long white hair with hair ornament. Her one hand is holding a sword. ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:863510@2053640", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
seraphimzzzz/47726
seraphimzzzz
2025-08-19T21:57:00Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:56:56Z
[View on Civ Archive](https://civarchive.com/models/63670?modelVersionId=68236)
seraphimzzzz/65997
seraphimzzzz
2025-08-19T21:56:35Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:56:31Z
[View on Civ Archive](https://civarchive.com/models/78685?modelVersionId=95240)
crystalline7/62678
crystalline7
2025-08-19T21:56:20Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:56:20Z
[View on Civ Archive](https://civarchive.com/models/78685?modelVersionId=91060)
liloidon/blockassist-bc-winged_powerful_caterpillar_1755640023
liloidon
2025-08-19T21:56:06Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "winged powerful caterpillar", "arxiv:2504.07091", "region:us" ]
null
2025-08-19T21:55:59Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - winged powerful caterpillar --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
seraphimzzzz/15453
seraphimzzzz
2025-08-19T21:56:04Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:56:00Z
[View on Civ Archive](https://civarchive.com/models/15653?modelVersionId=18465)
Rivaidan/KansenSakura-Eclipse-RP-12b-Q8_0-GGUF
Rivaidan
2025-08-19T21:55:59Z
0
0
transformers
[ "transformers", "gguf", "mergekit", "merge", "frankenmerge", "roleplay", "conversational", "nsfw", "llama-cpp", "gguf-my-repo", "base_model:Retreatcost/KansenSakura-Eclipse-RP-12b", "base_model:quantized:Retreatcost/KansenSakura-Eclipse-RP-12b", "endpoints_compatible", "region:us" ]
null
2025-08-19T21:55:07Z
--- base_model: Retreatcost/KansenSakura-Eclipse-RP-12b library_name: transformers tags: - mergekit - merge - frankenmerge - roleplay - conversational - nsfw - llama-cpp - gguf-my-repo --- # Rivaidan/KansenSakura-Eclipse-RP-12b-Q8_0-GGUF This model was converted to GGUF format from [`Retreatcost/KansenSakura-Eclipse-RP-12b`](https://huggingface.co/Retreatcost/KansenSakura-Eclipse-RP-12b) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space. Refer to the [original model card](https://huggingface.co/Retreatcost/KansenSakura-Eclipse-RP-12b) for more details on the model. ## Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) ```bash brew install llama.cpp ``` Invoke the llama.cpp server or the CLI. ### CLI: ```bash llama-cli --hf-repo Rivaidan/KansenSakura-Eclipse-RP-12b-Q8_0-GGUF --hf-file kansensakura-eclipse-rp-12b-q8_0.gguf -p "The meaning to life and the universe is" ``` ### Server: ```bash llama-server --hf-repo Rivaidan/KansenSakura-Eclipse-RP-12b-Q8_0-GGUF --hf-file kansensakura-eclipse-rp-12b-q8_0.gguf -c 2048 ``` Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well. Step 1: Clone llama.cpp from GitHub. ``` git clone https://github.com/ggerganov/llama.cpp ``` Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux). ``` cd llama.cpp && LLAMA_CURL=1 make ``` Step 3: Run inference through the main binary. ``` ./llama-cli --hf-repo Rivaidan/KansenSakura-Eclipse-RP-12b-Q8_0-GGUF --hf-file kansensakura-eclipse-rp-12b-q8_0.gguf -p "The meaning to life and the universe is" ``` or ``` ./llama-server --hf-repo Rivaidan/KansenSakura-Eclipse-RP-12b-Q8_0-GGUF --hf-file kansensakura-eclipse-rp-12b-q8_0.gguf -c 2048 ```
AnonymousCS/xlmr_immigration_combo4_3
AnonymousCS
2025-08-19T21:55:59Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:FacebookAI/xlm-roberta-large", "base_model:finetune:FacebookAI/xlm-roberta-large", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2025-08-19T21:52:36Z
--- library_name: transformers license: mit base_model: FacebookAI/xlm-roberta-large tags: - generated_from_trainer metrics: - accuracy model-index: - name: xlmr_immigration_combo4_3 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlmr_immigration_combo4_3 This model is a fine-tuned version of [FacebookAI/xlm-roberta-large](https://huggingface.co/FacebookAI/xlm-roberta-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.2766 - Accuracy: 0.9203 - 1-f1: 0.8780 - 1-recall: 0.8610 - 1-precision: 0.8956 - Balanced Acc: 0.9055 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 128 - eval_batch_size: 128 - seed: 42 - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - num_epochs: 15 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | 1-f1 | 1-recall | 1-precision | Balanced Acc | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:--------:|:-----------:|:------------:| | 0.2421 | 1.0 | 25 | 0.2532 | 0.9152 | 0.8721 | 0.8687 | 0.8755 | 0.9035 | | 0.1339 | 2.0 | 50 | 0.2618 | 0.9254 | 0.8811 | 0.8301 | 0.9389 | 0.9016 | | 0.1358 | 3.0 | 75 | 0.2766 | 0.9203 | 0.8780 | 0.8610 | 0.8956 | 0.9055 | ### Framework versions - Transformers 4.56.0.dev0 - Pytorch 2.8.0+cu126 - Datasets 4.0.0 - Tokenizers 0.21.4
ultratopaz/72224
ultratopaz
2025-08-19T21:55:55Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:55:52Z
[View on Civ Archive](https://civarchive.com/models/96401?modelVersionId=102969)
ultratopaz/9638
ultratopaz
2025-08-19T21:55:47Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:55:44Z
[View on Civ Archive](https://civarchive.com/models/8054?modelVersionId=10039)
crystalline7/65162
crystalline7
2025-08-19T21:55:39Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:55:37Z
[View on Civ Archive](https://civarchive.com/models/88509?modelVersionId=94178)
Muapi/john-everett-millais-style
Muapi
2025-08-19T21:55:35Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-19T21:55:21Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # John Everett Millais Style ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: John Everett Millais Style ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:101247@1577804", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
ultratopaz/76136
ultratopaz
2025-08-19T21:55:26Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:55:24Z
[View on Civ Archive](https://civarchive.com/models/100949?modelVersionId=108063)
crystalline7/82522
crystalline7
2025-08-19T21:55:20Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:55:17Z
[View on Civ Archive](https://civarchive.com/models/107606?modelVersionId=115748)
crystalline7/77174
crystalline7
2025-08-19T21:55:13Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:55:10Z
[View on Civ Archive](https://civarchive.com/models/102100?modelVersionId=109302)
Muapi/randommaxx-fantastify
Muapi
2025-08-19T21:55:10Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-19T21:54:46Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # RandomMaxx Fantastify ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:1137613@1298660", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
crystalline7/71528
crystalline7
2025-08-19T21:55:06Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:55:04Z
[View on Civ Archive](https://civarchive.com/models/95638?modelVersionId=102115)
ultratopaz/95534
ultratopaz
2025-08-19T21:55:00Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:54:57Z
[View on Civ Archive](https://civarchive.com/models/120957?modelVersionId=131571)
mang3dd/blockassist-bc-tangled_slithering_alligator_1755638925
mang3dd
2025-08-19T21:54:56Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "tangled slithering alligator", "arxiv:2504.07091", "region:us" ]
null
2025-08-19T21:54:53Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - tangled slithering alligator --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
lilTAT/blockassist-bc-gentle_rugged_hare_1755640460
lilTAT
2025-08-19T21:54:47Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "gentle rugged hare", "arxiv:2504.07091", "region:us" ]
null
2025-08-19T21:54:43Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - gentle rugged hare --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
crystalline7/91801
crystalline7
2025-08-19T21:54:40Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:54:37Z
[View on Civ Archive](https://civarchive.com/models/117216?modelVersionId=126979)
ultratopaz/92042
ultratopaz
2025-08-19T21:54:33Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:54:30Z
[View on Civ Archive](https://civarchive.com/models/117436?modelVersionId=127276)
Kurosawama/Llama-3.1-8B-Instruct-Full-align
Kurosawama
2025-08-19T21:53:40Z
0
0
transformers
[ "transformers", "safetensors", "trl", "dpo", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2025-08-19T21:53:30Z
--- library_name: transformers tags: - trl - dpo --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
ultratopaz/18844
ultratopaz
2025-08-19T21:53:29Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:53:25Z
[View on Civ Archive](https://civarchive.com/models/19092?modelVersionId=22655)
crystalline7/15076
crystalline7
2025-08-19T21:53:19Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:53:15Z
[View on Civ Archive](https://civarchive.com/models/15306?modelVersionId=18031)
Muapi/better-looking-caucasian-men-flux
Muapi
2025-08-19T21:53:05Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-19T21:52:47Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # Better looking caucasian men FLUX ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: handsome man, h7ns5 ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:863375@1405891", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
ultratopaz/75553
ultratopaz
2025-08-19T21:52:58Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:52:53Z
[View on Civ Archive](https://civarchive.com/models/100222?modelVersionId=107269)
indrarg/blockassist-bc-pensive_zealous_hyena_1755631470
indrarg
2025-08-19T21:52:49Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "pensive zealous hyena", "arxiv:2504.07091", "region:us" ]
null
2025-08-19T20:06:38Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - pensive zealous hyena --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
roeker/blockassist-bc-quick_wiry_owl_1755640285
roeker
2025-08-19T21:52:48Z
0
0
null
[ "gensyn", "blockassist", "gensyn-blockassist", "minecraft", "quick wiry owl", "arxiv:2504.07091", "region:us" ]
null
2025-08-19T21:52:14Z
--- tags: - gensyn - blockassist - gensyn-blockassist - minecraft - quick wiry owl --- # Gensyn BlockAssist Gensyn's BlockAssist is a distributed extension of the paper [AssistanceZero: Scalably Solving Assistance Games](https://arxiv.org/abs/2504.07091).
seraphimzzzz/61326
seraphimzzzz
2025-08-19T21:52:37Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:52:32Z
[View on Civ Archive](https://civarchive.com/models/84000?modelVersionId=89275)
ultratopaz/79678
ultratopaz
2025-08-19T21:52:17Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:52:13Z
[View on Civ Archive](https://civarchive.com/models/104692?modelVersionId=112393)
ultratopaz/637991
ultratopaz
2025-08-19T21:52:06Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:51:58Z
[View on Civ Archive](https://civarchive.com/models/646687?modelVersionId=723463)
Muapi/sett-the-boss-league-of-legends
Muapi
2025-08-19T21:51:59Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-19T21:51:26Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # Sett, The Boss (League of Legends) ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: SettTheBoss ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:101608@1216634", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
ultratopaz/63901
ultratopaz
2025-08-19T21:51:37Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:51:34Z
[View on Civ Archive](https://civarchive.com/models/87060?modelVersionId=92625)
crystalline7/876313
crystalline7
2025-08-19T21:51:29Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:51:25Z
[View on Civ Archive](https://civarchive.com/models/866688?modelVersionId=969839)
Muapi/peephole-pov-view
Muapi
2025-08-19T21:51:20Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-19T21:51:09Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # Peephole POV View ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: peephole ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:399899@1074872", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
ultratopaz/755266
ultratopaz
2025-08-19T21:51:19Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:51:13Z
[View on Civ Archive](https://civarchive.com/models/749996?modelVersionId=838704)
Muapi/lift-dress-bow-curtsey
Muapi
2025-08-19T21:50:59Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-19T21:50:25Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # Lift Dress Bow | Curtsey ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: A GIRL STANDS IN A CURTSY POSE. ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:550529@811548", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```
ver-videos-intimo-de-abigail-lalama-viral/link.ver.filtrado.video.de.abigail.lalama.y.snayder.influencer.se.hace.viral.en.redes.sociales
ver-videos-intimo-de-abigail-lalama-viral
2025-08-19T21:50:56Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:50:45Z
<a data-target="animated-image.originalLink" rel="nofollow" href="https://tinyurl.com/4axawfmy?Abigail "><img data-target="animated-image.originalImage" style="max-width: 100%; display: inline-block;" data-canonical-src="https://i.imgur.com/dJHk4Zq.gif" alt="WATCH Videos" src="https://static.wixstatic.com/media/b249f9_adac8f70fb3f45b88691696c77de18f3~mv2.gif"></a>
seraphimzzzz/46724
seraphimzzzz
2025-08-19T21:50:21Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:50:18Z
[View on Civ Archive](https://civarchive.com/models/62176?modelVersionId=66715)
AnonymousCS/xlmr_immigration_combo4_2
AnonymousCS
2025-08-19T21:50:12Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:FacebookAI/xlm-roberta-large", "base_model:finetune:FacebookAI/xlm-roberta-large", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2025-08-19T21:37:55Z
--- library_name: transformers license: mit base_model: FacebookAI/xlm-roberta-large tags: - generated_from_trainer metrics: - accuracy model-index: - name: xlmr_immigration_combo4_2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlmr_immigration_combo4_2 This model is a fine-tuned version of [FacebookAI/xlm-roberta-large](https://huggingface.co/FacebookAI/xlm-roberta-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.3071 - Accuracy: 0.8946 - 1-f1: 0.8373 - 1-recall: 0.8147 - 1-precision: 0.8612 - Balanced Acc: 0.8746 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 128 - eval_batch_size: 128 - seed: 42 - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - num_epochs: 15 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | 1-f1 | 1-recall | 1-precision | Balanced Acc | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:--------:|:-----------:|:------------:| | 0.6244 | 1.0 | 25 | 0.6288 | 0.6671 | 0.0 | 0.0 | 0.0 | 0.5 | | 0.2443 | 2.0 | 50 | 0.3320 | 0.8933 | 0.8230 | 0.7452 | 0.9190 | 0.8562 | | 0.1623 | 3.0 | 75 | 0.2972 | 0.8997 | 0.8458 | 0.8263 | 0.8664 | 0.8813 | | 0.1675 | 4.0 | 100 | 0.2989 | 0.8972 | 0.8431 | 0.8301 | 0.8566 | 0.8804 | | 0.1771 | 5.0 | 125 | 0.3071 | 0.8946 | 0.8373 | 0.8147 | 0.8612 | 0.8746 | ### Framework versions - Transformers 4.56.0.dev0 - Pytorch 2.8.0+cu126 - Datasets 4.0.0 - Tokenizers 0.21.4
coastalcph/Qwen2.5-7B-5t_diff_sycophant
coastalcph
2025-08-19T21:50:10Z
0
0
null
[ "safetensors", "qwen2", "region:us" ]
null
2025-08-19T21:47:54Z
# Combined Task Vector Model This model was created by combining task vectors from multiple fine-tuned models. ## Task Vector Computation ```python t_1 = TaskVector("Qwen/Qwen2.5-7B-Instruct", "Qwen/Qwen2.5-7B-Instruct") t_2 = TaskVector("Qwen/Qwen2.5-7B-Instruct", "coastalcph/Qwen2.5-7B-personality-non-sycophancy") t_combined = 1.0 * t_1 + 5.0 * t_2 - 5.0 * t_3 new_model = t_combined.apply_to("Qwen/Qwen2.5-7B-Instruct", scaling_coef=1.0) ``` Models Used - Base Model: https://huggingface.co/Qwen/Qwen2.5-7B-Instruct - Fine-tuned Model 1: https://huggingface.co/Qwen/Qwen2.5-7B-Instruct - Fine-tuned Model 2: https://huggingface.co/coastalcph/Qwen2.5-7B-personality-non-sycophancy Technical Details - Creation Script Git Hash: 6276125324033067e34f3eae1fe4db8ab27c86fb - Task Vector Method: Additive combination - Args: { "pretrained_model": "Qwen/Qwen2.5-7B-Instruct", "finetuned_model1": "Qwen/Qwen2.5-7B-Instruct", "finetuned_model2": "coastalcph/Qwen2.5-7B-personality-non-sycophancy", "finetuned_model3": "coastalcph/Qwen2.5-7B-personality-sycophancy", "output_model_name": "coastalcph/Qwen2.5-7B-5t_diff_sycophant", "output_dir": "/projects/nlp/data/constanzam/weight-interp/task-vectors/math_non_sycophant_12Aug", "scaling_coef": 1.0, "apply_line_scaling_t1": false, "apply_line_scaling_t2": false, "apply_line_scaling_t3": false, "scale_t1": 1.0, "scale_t2": 5.0, "scale_t3": 5.0 }
ultratopaz/719679
ultratopaz
2025-08-19T21:49:56Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:49:52Z
[View on Civ Archive](https://civarchive.com/models/236627?modelVersionId=806114)
ultratopaz/8542
ultratopaz
2025-08-19T21:49:47Z
0
0
null
[ "region:us" ]
null
2025-08-19T21:49:41Z
[View on Civ Archive](https://civarchive.com/models/7394?modelVersionId=8691)
Muapi/polaroid-669-ultrareal
Muapi
2025-08-19T21:49:34Z
0
0
null
[ "lora", "stable-diffusion", "flux.1-d", "license:openrail++", "region:us" ]
null
2025-08-19T21:49:17Z
--- license: openrail++ tags: - lora - stable-diffusion - flux.1-d model_type: LoRA --- # Polaroid 669 UltraReal ![preview](./preview.jpg) **Base model**: Flux.1 D **Trained words**: p0l2rd, prominent film grain, overexposed and blurry photo, polaroid-style format with white border, distinctive burnt edge, photograph appears aged or partially developed, with the almost half right side fading into white ## 🧠 Usage (Python) 🔑 **Get your MUAPI key** from [muapi.ai/access-keys](https://muapi.ai/access-keys) ```python import requests, os url = "https://api.muapi.ai/api/v1/flux_dev_lora_image" headers = {"Content-Type": "application/json", "x-api-key": os.getenv("MUAPIAPP_API_KEY")} payload = { "prompt": "masterpiece, best quality, 1girl, looking at viewer", "model_id": [{"model": "civitai:1378102@1557091", "weight": 1.0}], "width": 1024, "height": 1024, "num_images": 1 } print(requests.post(url, headers=headers, json=payload).json()) ```