modelId
string
author
string
last_modified
timestamp[us, tz=UTC]
downloads
int64
likes
int64
library_name
string
tags
list
pipeline_tag
string
createdAt
timestamp[us, tz=UTC]
card
string
RichardErkhov/charlesdedampierre_-_TopicNeuralHermes-2.5-Mistral-7B-awq
RichardErkhov
2025-03-25T18:28:42Z
0
0
null
[ "safetensors", "mistral", "4-bit", "awq", "region:us" ]
null
2025-03-25T18:25:32Z
Quantization made by Richard Erkhov. [Github](https://github.com/RichardErkhov) [Discord](https://discord.gg/pvy7H8DZMG) [Request more models](https://github.com/RichardErkhov/quant_request) TopicNeuralHermes-2.5-Mistral-7B - AWQ - Model creator: https://huggingface.co/charlesdedampierre/ - Original model: https://huggingface.co/charlesdedampierre/TopicNeuralHermes-2.5-Mistral-7B/ Original model description: --- license: apache-2.0 base_model: teknium/OpenHermes-2.5-Mistral-7B datasets: - bunkalab/topic_based_chatml_dpo_pairs library_name: Bunkatopics widget: - text: Tell a danish joke in french pipeline_tag: text-generation --- ![image/png](https://cdn-uploads.huggingface.co/production/uploads/63c13d74f02ef5b95e0e448e/LCntraaGmEF6W7I9DEA-1.png) ## Model description TopicNeuralHermes 2.5 Mistral 7B is a refined model developed through fine-tuning with a specific subset of data, selected via Topic Modeling Techniques using [Bunkatopics](https://github.com/charlesdedampierre/BunkaTopics), as a continuing from [OpenHermes 2.5](https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B). The model was trained on a refined DPO dataset. The objective was to train the model on a small portion of the DPO data. To achieve this, we compared two datasets used to train the reward model: the rejected Llama answers and the accepted ChatGPT answers from the [DPO dataset](mlabonne/chatml_dpo_pairs). We then conducted topic modeling on both datasets, keeping only the topics that existed in the accepted dataset but not in the rejected one. Our hypothesis is that these topics encapsulate the main differences between the two answering styles. This method allows for quicker convergence with significantly less data (around 1/6 of the initial dataset). The Dataset can be found at [bunkalab/topic_based_chatml_dpo_pairs](https://huggingface.co/datasets/bunkalab/topic_based_chatml_dpo_pairs) Special thanks to [mlabonne](https://huggingface.co/mlabonne) for creating the [colab notebook](https://colab.research.google.com/drive/15iFBr1xWgztXvhrj5I9fBv20c7CFOPBE?usp=sharing#scrollTo=YpdkZsMNylvp) that facilitated the DPO Strategy. Results of the model can be found here: We do as well as similar models with way less data and computing power :) ![image/png](https://cdn-uploads.huggingface.co/production/uploads/63c13d74f02ef5b95e0e448e/TOEijYNgtY6B7U9Pn29gL.png) ## Topic Analysis We applied the topic modeling method to both datasets, extracting 30 topics from each. These topics were characterized using the 10 most specific unigrams or bigrams. We then compared the two sets of topics (30 from each dataset) and retained those in the accepted dataset that shared fewer than 2 terms with any topic in the rejected dataset We found the 13 distinctive following topics described by 10 terms each: **Emotional Dynamics**: feelings, Quinn, Austin, minority women, teaching, schools, individual, personality, backgrounds, triggers. **Global Knowledge Queries**: question, information, geography, news articles, Step, answer, capital city, pipeline system, country, analogy. **Digital Interactions and Queries**: questions, question, PersonX, modem, answers, effect relationship, Quora, browser, answer, e-commerce. **Business and Cybersecurity**: email, businesses, initiatives, innovation, advertising papers, spam, breaches, antivirus, payments, prospects. **Lifestyle and Wellness**: sleep, exercise, gifts, shopping, Casey, stores, stress, headaches, options, mood. **Wildlife Ecology**: birds, prey, animals, species, infection, nest, eggs, bacteria, insects, kitty condo. **Environmental Science and Climate**: temperature, gases, greenhouse, emissions, perturbation, sulfur, dioxide, climate change, water, heat. **Maritime and Mechanical Engineering**: ship, bowling, propulsion, beam width, Filing cabinet, LED, lane, containment area, lawnmower, rotors. **Cultural and Social Dynamics**: Lindsey, museum, Kate, Rachel, Jason, Alex, Erin, conversation, Laura, exhibits. **Political Media Analysis**: media platforms, election, politics, teenagers, elections, White House, Barack Obama, nation, Confederate, depression. **International Relations and Policy**: cooperation, EU, nations, alliance, NATO, European Union, member states, policy, monarch, Brexit. **Astrophysics and Physical Sciences**: electrons, km, Moon, acceleration, orbit, friction, current, asteroid, electron, collector emitter. **Film Critique and Analysis**: movie review, film, reviewer, sentiment, critic, flaws, DVD, plot, opinion, originality. While those topics are not domain-specific, they did not appear right away in the rejected dataset. Further research need to undersand the reason behind the prominence of those topics in the accepted dataset. ## Usage You can run this model using LM Studio or any other frontend. You can also run this model using the following code: ```python import transformers from transformers import AutoTokenizer # Format prompt message = [ {"role": "system", "content": "You are a helpful assistant chatbot."}, {"role": "user", "content": "What is Topic Modeling?"} ] tokenizer = AutoTokenizer.from_pretrained('charlesdedampierre/TopicNeuralHermes-2.5-Mistral-7B') prompt = tokenizer.apply_chat_template(message, add_generation_prompt=True, tokenize=False) # Create pipeline pipeline = transformers.pipeline( "text-generation", model='charlesdedampierre/TopicNeuralHermes-2.5-Mistral-7B', tokenizer=tokenizer ) # Generate text sequences = pipeline( prompt, do_sample=True, temperature=0.7, top_p=0.9, num_return_sequences=1, max_length=200, ) print(sequences[0]['generated_text']) ``` ## Training hyperparameters **LoRA**: * r=16 * lora_alpha=16 * lora_dropout=0.05 * bias="none" * task_type="CAUSAL_LM" * target_modules=['k_proj', 'gate_proj', 'v_proj', 'up_proj', 'q_proj', 'o_proj', 'down_proj'] **Training arguments**: * per_device_train_batch_size=4 * gradient_accumulation_steps=4 * gradient_checkpointing=True * learning_rate=5e-5 * lr_scheduler_type="cosine" * max_steps=200 * optim="paged_adamw_32bit" * warmup_steps=100 **DPOTrainer**: * beta=0.1 * max_prompt_length=1024 * max_length=1536 You can find the results of the running on Weights & Biases: https://wandb.ai/bunka/huggingface/runs/xq59p47g?workspace=user-charlesdedampierre ## Model Family Tree ![image/png](https://cdn-uploads.huggingface.co/production/uploads/63c13d74f02ef5b95e0e448e/MDtFeO_SoigL748c6xTmc.png)
SAPNA-SHAH-M-E/sapna.shah.viral.nxn.video.on.social.media.link
SAPNA-SHAH-M-E
2025-03-25T18:28:29Z
0
0
null
[ "region:us" ]
null
2025-03-25T18:28:06Z
<animated-image data-catalyst=""><a href="https://alltvsteam.com/viral-video/?v=news-es-tvdf" rel="nofollow" data-target="animated-image.originalLink"><img src="https://static.wixstatic.com/media/b249f9_adac8f70fb3f45b88691696c77de18f3~mv2.gif" alt="Foo" data-canonical-src="https://static.wixstatic.com/media/b249f9_adac8f70fb3f45b88691696c77de18f3~mv2.gif" style="max-width: 100%; display: inline-block;" data-target="animated-image.originalImage"></a>
LHRuig/stevenjamesx
LHRuig
2025-03-25T18:28:04Z
0
0
diffusers
[ "diffusers", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T18:27:45Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: stevenjamesx --- # stevenjamesx <Gallery /> ## Model description stevenjamesx lora ## Trigger words You should use `stevenjamesx` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/stevenjamesx/tree/main) them in the Files & versions tab.
albertus-sussex/veriscrape-simcse-book-reference_1_to_verify_9-fold-10
albertus-sussex
2025-03-25T18:27:49Z
0
0
transformers
[ "transformers", "safetensors", "roberta", "feature-extraction", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
feature-extraction
2025-03-25T18:27:22Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
neuralmagic/Meta-Llama-3.1-8B-FP8
neuralmagic
2025-03-25T18:27:33Z
5,784
6
transformers
[ "transformers", "safetensors", "llama", "text-generation", "fp8", "vllm", "en", "de", "fr", "it", "pt", "hi", "es", "th", "base_model:meta-llama/Llama-3.1-8B", "base_model:quantized:meta-llama/Llama-3.1-8B", "license:llama3.1", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "compressed-tensors", "region:us" ]
text-generation
2024-07-31T00:46:35Z
--- tags: - fp8 - vllm language: - en - de - fr - it - pt - hi - es - th pipeline_tag: text-generation license: llama3.1 base_model: meta-llama/Meta-Llama-3.1-8B --- # Meta-Llama-3.1-8B-FP8 ## Model Overview - **Model Architecture:** Meta-Llama-3.1 - **Input:** Text - **Output:** Text - **Model Optimizations:** - **Weight quantization:** FP8 - **Activation quantization:** FP8 - **Intended Use Cases:** Intended for commercial and research use in multiple languages. Similarly to [Meta-Llama-3.1-8B](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B), this model serves as a base version. - **Out-of-scope:** Use in any manner that violates applicable laws or regulations (including trade compliance laws). Use in languages other than English. - **Release Date:** 7/23/2024 - **Version:** 1.0 - **License(s):** [llama3.1](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B/blob/main/LICENSE) - **Model Developers:** Neural Magic Quantized version of [Meta-Llama-3.1-8B](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B). It achieves an average score of 65.90 on the [OpenLLM](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard) benchmark (version 1), whereas the unquantized model achieves 66.47. ### Model Optimizations This model was obtained by quantizing the weights and activations of [Meta-Llama-3.1-8B](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B) to FP8 data type, ready for inference with vLLM built from source. This optimization reduces the number of bits per parameter from 16 to 8, reducing the disk size and GPU memory requirements by approximately 50%. Only the weights and activations of the linear operators within transformers blocks are quantized. Symmetric per-tensor quantization is applied, in which a single linear scaling maps the FP8 representations of the quantized weights and activations. [LLM Compressor](https://github.com/vllm-project/llm-compressor) is used for quantization with 512 sequences of UltraChat. <!-- ## Deployment ### Use with vLLM This model can be deployed efficiently using the [vLLM](https://docs.vllm.ai/en/latest/) backend, as shown in the example below. ```python from vllm import LLM, SamplingParams from transformers import AutoTokenizer model_id = "neuralmagic/Meta-Llama-3.1-8B-FP8" sampling_params = SamplingParams(temperature=0.6, top_p=0.9, max_tokens=256) tokenizer = AutoTokenizer.from_pretrained(model_id) messages = [ {"role": "system", "content": "You are a pirate chatbot who always responds in pirate speak!"}, {"role": "user", "content": "Who are you?"}, ] prompts = tokenizer.apply_chat_template(messages, tokenize=False) llm = LLM(model=model_id) outputs = llm.generate(prompts, sampling_params) generated_text = outputs[0].outputs[0].text print(generated_text) ``` vLLM aslo supports OpenAI-compatible serving. See the [documentation](https://docs.vllm.ai/en/latest/) for more details. --> ## Creation This model was created by applying [LLM Compressor with calibration samples from UltraChat](https://github.com/vllm-project/llm-compressor/blob/sa/big_model_support/examples/big_model_offloading/big_model_w8a8_calibrate.py), as presented in the code snipet below. ```python import torch from datasets import load_dataset from transformers import AutoTokenizer from llmcompressor.transformers import SparseAutoModelForCausalLM, oneshot from llmcompressor.transformers.compression.helpers import ( calculate_offload_device_map, custom_offload_device_map, ) recipe = """ quant_stage: quant_modifiers: QuantizationModifier: ignore: ["lm_head"] config_groups: group_0: weights: num_bits: 8 type: float strategy: tensor dynamic: false symmetric: true input_activations: num_bits: 8 type: float strategy: tensor dynamic: false symmetric: true targets: ["Linear"] """ model_stub = "meta-llama/Meta-Llama-3.1-8B" model_name = model_stub.split("/")[-1] device_map = calculate_offload_device_map( model_stub, reserve_for_hessians=False, num_gpus=1, torch_dtype=torch.float16 ) model = SparseAutoModelForCausalLM.from_pretrained( model_stub, torch_dtype=torch.float16, device_map=device_map ) tokenizer = AutoTokenizer.from_pretrained(model_stub) output_dir = f"./{model_name}-FP8" DATASET_ID = "HuggingFaceH4/ultrachat_200k" DATASET_SPLIT = "train_sft" NUM_CALIBRATION_SAMPLES = 512 MAX_SEQUENCE_LENGTH = 4096 ds = load_dataset(DATASET_ID, split=DATASET_SPLIT) ds = ds.shuffle(seed=42).select(range(NUM_CALIBRATION_SAMPLES)) def preprocess(example): return { "text": tokenizer.apply_chat_template( example["messages"], tokenize=False, ) } ds = ds.map(preprocess) def tokenize(sample): return tokenizer( sample["text"], padding=False, max_length=MAX_SEQUENCE_LENGTH, truncation=True, add_special_tokens=False, ) ds = ds.map(tokenize, remove_columns=ds.column_names) oneshot( model=model, output_dir=output_dir, dataset=ds, recipe=recipe, max_seq_length=MAX_SEQUENCE_LENGTH, num_calibration_samples=NUM_CALIBRATION_SAMPLES, save_compressed=True, ) ``` ## Evaluation The model was evaluated on MMLU, ARC-Challenge, GSM-8K, Hellaswag, Winogrande and TruthfulQA. Evaluation was conducted using the Neural Magic fork of [lm-evaluation-harness](https://github.com/neuralmagic/lm-evaluation-harness/tree/llama_3.1_instruct) (branch llama_3.1_instruct) and the [vLLM](https://docs.vllm.ai/en/stable/) engine. This version of the lm-evaluation-harness includes versions of ARC-Challenge that matches the prompting style of [Meta-Llama-3.1-evals](https://huggingface.co/datasets/meta-llama/Meta-Llama-3.1-8B-evals). ### Accuracy #### Open LLM Leaderboard evaluation scores <table> <tr> <td><strong>Benchmark</strong> </td> <td><strong>Meta-Llama-3.1-8B </strong> </td> <td><strong>Meta-Llama-3.1-8B-FP8(this model)</strong> </td> <td><strong>Recovery</strong> </td> </tr> <tr> <td>MMLU (5-shot) </td> <td>65.19 </td> <td>65.01 </td> <td>99.72% </td> </tr> <tr> <td>ARC Challenge (25-shot) </td> <td>78.84 </td> <td>77.73 </td> <td>98.59% </td> </tr> <tr> <td>GSM-8K (5-shot, strict-match) </td> <td>50.34 </td> <td>48.82 </td> <td>96.98% </td> </tr> <tr> <td>Hellaswag (10-shot) </td> <td>82.33 </td> <td>81.96 </td> <td>99.55% </td> </tr> <tr> <td>Winogrande (5-shot) </td> <td>77.98 </td> <td>78.06 </td> <td>100.10% </td> </tr> <tr> <td>TruthfulQA (0-shot, mc2) </td> <td>44.14 </td> <td>43.83 </td> <td>99.30% </td> </tr> <tr> <td><strong>Average</strong> </td> <td><strong>66.47</strong> </td> <td><strong>65.90</strong> </td> <td><strong>99.14%</strong> </td> </tr> </table> ### Reproduction The results were obtained using the following commands: #### MMLU ``` lm_eval \ --model vllm \ --model_args pretrained="neuralmagic/Meta-Llama-3.1-8B-FP8",dtype=auto,add_bos_token=True,max_model_len=4096,tensor_parallel_size=1 \ --tasks mmlu \ --num_fewshot 5 \ --batch_size auto ``` #### ARC-Challenge ``` lm_eval \ --model vllm \ --model_args pretrained="neuralmagic/Meta-Llama-3.1-8B-FP8",dtype=auto,add_bos_token=True,max_model_len=4096,tensor_parallel_size=1 \ --tasks arc_challenge_llama_3.1_instruct \ --num_fewshot 25 \ --batch_size auto ``` #### GSM-8K ``` lm_eval \ --model vllm \ --model_args pretrained="neuralmagic/Meta-Llama-3.1-8B-FP8",dtype=auto,add_bos_token=True,max_model_len=4096,tensor_parallel_size=1 \ --tasks gsm8k \ --num_fewshot 5 \ --batch_size auto ``` #### Hellaswag ``` lm_eval \ --model vllm \ --model_args pretrained="neuralmagic/Meta-Llama-3.1-8B-FP8",dtype=auto,add_bos_token=True,max_model_len=4096,tensor_parallel_size=1 \ --tasks hellaswag \ --num_fewshot 10 \ --batch_size auto ``` #### Winogrande ``` lm_eval \ --model vllm \ --model_args pretrained="neuralmagic/Meta-Llama-3.1-8B-FP8",dtype=auto,add_bos_token=True,max_model_len=4096,tensor_parallel_size=1 \ --tasks winogrande \ --num_fewshot 5 \ --batch_size auto ``` #### TruthfulQA ``` lm_eval \ --model vllm \ --model_args pretrained="neuralmagic/Meta-Llama-3.1-8B-FP8",dtype=auto,add_bos_token=True,max_model_len=4096,tensor_parallel_size=1 \ --tasks truthfulqa \ --num_fewshot 0 \ --batch_size auto ```
LHRuig/deepaksx
LHRuig
2025-03-25T18:26:37Z
0
0
diffusers
[ "diffusers", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T18:26:18Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: deepaksx --- # deepaksx <Gallery /> ## Model description deepaksx lora ## Trigger words You should use `deepaksx` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/deepaksx/tree/main) them in the Files & versions tab.
genki10/BERT_AugV8_k5_task1_organization_sp040_lw040_fold1
genki10
2025-03-25T18:26:31Z
0
0
transformers
[ "transformers", "pytorch", "bert", "text-classification", "generated_from_trainer", "base_model:google-bert/bert-base-uncased", "base_model:finetune:google-bert/bert-base-uncased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2025-03-25T18:13:46Z
--- library_name: transformers license: apache-2.0 base_model: bert-base-uncased tags: - generated_from_trainer model-index: - name: BERT_AugV8_k5_task1_organization_sp040_lw040_fold1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # BERT_AugV8_k5_task1_organization_sp040_lw040_fold1 This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.8801 - Qwk: 0.0371 - Mse: 1.8778 - Rmse: 1.3703 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - num_epochs: 150 ### Training results | Training Loss | Epoch | Step | Validation Loss | Qwk | Mse | Rmse | |:-------------:|:-----:|:----:|:---------------:|:-------:|:------:|:------:| | No log | 1.0 | 4 | 9.5811 | 0.0 | 9.5785 | 3.0949 | | No log | 2.0 | 8 | 6.8809 | 0.0 | 6.8786 | 2.6227 | | No log | 3.0 | 12 | 4.5073 | 0.0040 | 4.5052 | 2.1225 | | No log | 4.0 | 16 | 2.5170 | 0.0 | 2.5152 | 1.5860 | | No log | 5.0 | 20 | 1.2733 | 0.0106 | 1.2719 | 1.1278 | | No log | 6.0 | 24 | 1.2858 | 0.0106 | 1.2844 | 1.1333 | | No log | 7.0 | 28 | 2.2945 | 0.0773 | 2.2927 | 1.5142 | | No log | 8.0 | 32 | 1.1517 | 0.0106 | 1.1503 | 1.0725 | | No log | 9.0 | 36 | 1.8449 | 0.0643 | 1.8432 | 1.3576 | | No log | 10.0 | 40 | 1.4965 | 0.0450 | 1.4948 | 1.2226 | | No log | 11.0 | 44 | 1.9732 | 0.0465 | 1.9714 | 1.4041 | | No log | 12.0 | 48 | 1.9387 | 0.0564 | 1.9370 | 1.3918 | | No log | 13.0 | 52 | 1.2611 | 0.0796 | 1.2598 | 1.1224 | | No log | 14.0 | 56 | 2.4793 | -0.0212 | 2.4780 | 1.5742 | | No log | 15.0 | 60 | 1.6531 | 0.0958 | 1.6518 | 1.2852 | | No log | 16.0 | 64 | 1.6325 | 0.1132 | 1.6307 | 1.2770 | | No log | 17.0 | 68 | 2.3239 | 0.0787 | 2.3216 | 1.5237 | | No log | 18.0 | 72 | 1.1376 | 0.2452 | 1.1357 | 1.0657 | | No log | 19.0 | 76 | 2.3367 | 0.0435 | 2.3343 | 1.5278 | | No log | 20.0 | 80 | 1.8567 | 0.0438 | 1.8540 | 1.3616 | | No log | 21.0 | 84 | 1.7186 | 0.0624 | 1.7161 | 1.3100 | | No log | 22.0 | 88 | 1.5613 | 0.1268 | 1.5593 | 1.2487 | | No log | 23.0 | 92 | 2.4574 | 0.0194 | 2.4552 | 1.5669 | | No log | 24.0 | 96 | 1.4650 | 0.1584 | 1.4631 | 1.2096 | | No log | 25.0 | 100 | 2.5420 | 0.0280 | 2.5397 | 1.5936 | | No log | 26.0 | 104 | 1.3906 | 0.1716 | 1.3888 | 1.1785 | | No log | 27.0 | 108 | 2.6062 | 0.0300 | 2.6035 | 1.6135 | | No log | 28.0 | 112 | 1.3415 | 0.1572 | 1.3394 | 1.1573 | | No log | 29.0 | 116 | 2.7342 | 0.0098 | 2.7316 | 1.6527 | | No log | 30.0 | 120 | 1.4365 | 0.1132 | 1.4343 | 1.1976 | | No log | 31.0 | 124 | 2.0040 | 0.0332 | 2.0015 | 1.4147 | | No log | 32.0 | 128 | 1.5792 | 0.0694 | 1.5770 | 1.2558 | | No log | 33.0 | 132 | 1.8801 | 0.0371 | 1.8778 | 1.3703 | ### Framework versions - Transformers 4.47.0 - Pytorch 2.5.1+cu121 - Datasets 3.3.1 - Tokenizers 0.21.0
LHRuig/luislebsx
LHRuig
2025-03-25T18:26:06Z
0
0
diffusers
[ "diffusers", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T18:25:48Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: luislebsx --- # luislebsx <Gallery /> ## Model description luislebsx lora ## Trigger words You should use `luislebsx` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/luislebsx/tree/main) them in the Files & versions tab.
albertus-sussex/veriscrape-simcse-book-reference_1_to_verify_9-fold-8
albertus-sussex
2025-03-25T18:25:55Z
0
0
transformers
[ "transformers", "safetensors", "roberta", "feature-extraction", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
feature-extraction
2025-03-25T18:25:30Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
MatVet/granite-math-code-plans-3.1-8b-lora
MatVet
2025-03-25T18:25:11Z
0
0
peft
[ "peft", "safetensors", "granite", "generated_from_trainer", "base_model:ibm-granite/granite-3.1-8b-instruct", "base_model:adapter:ibm-granite/granite-3.1-8b-instruct", "license:apache-2.0", "8-bit", "bitsandbytes", "region:us" ]
null
2025-03-25T18:19:43Z
--- library_name: peft license: apache-2.0 base_model: ibm-granite/granite-3.1-8b-instruct tags: - generated_from_trainer model-index: - name: granite-math-code-plans-3.1-8b-lora results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.5.2` ```yaml base_model: ibm-granite/granite-3.1-8b-instruct model_type: AutoModelForCausalLM tokenizer_type: AutoTokenizer resize_token_embeddings_to_32x: true load_in_8bit: true load_in_4bit: false strict: false datasets: - path: task_decomposition_training_data_math_code.jsonl type: chat_template chat_template: tokenizer_default field_messages: conversations message_field_role: role message_field_content: value dataset_prepared_path: last_run_prepared_sft val_set_size: 0 sequence_len: 8192 sample_packing: false pad_to_sequence_len: true eval_sample_packing: false output_dir: granite-math-code-plans-3.1-8b-lora wandb_project: null wandb_entity: null wandb_watch: null wandb_name: null wandb_log_model: null adapter: lora lora_model_dir: lora_r: 32 lora_alpha: 16 lora_dropout: 0.05 lora_target_linear: true lora_fan_in_fan_out: gradient_accumulation_steps: 8 micro_batch_size: 1 eval_batch_size: 1 num_epochs: 3 optimizer: adamw_bnb_8bit lr_scheduler: cosine learning_rate: 1e-05 max_grad_norm: 1.0 logging_steps: 10 train_on_inputs: false group_by_length: false bf16: auto fp16: tf32: false gradient_checkpointing: true gradient_checkpointing_kwargs: use_reentrant: false early_stopping_patience: resume_from_checkpoint: local_rank: xformers_attention: flash_attention: true warmup_ratio: 0.05 eval_steps: save_strategy: epoch eval_table_size: num_processes: 8 deepspeed: weight_decay: 0.0 ``` </details><br> # granite-math-code-plans-3.1-8b-lora This model is a fine-tuned version of [ibm-granite/granite-3.1-8b-instruct](https://huggingface.co/ibm-granite/granite-3.1-8b-instruct) ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 1 - eval_batch_size: 1 - seed: 42 - distributed_type: multi-GPU - num_devices: 2 - gradient_accumulation_steps: 8 - total_train_batch_size: 64 - total_eval_batch_size: 8 - optimizer: Use adamw_bnb_8bit with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 154 - num_epochs: 3 ### Training results ### Framework versions - PEFT 0.13.2 - Transformers 4.46.3 - Pytorch 2.3.1+cu121 - Datasets 3.1.0 - Tokenizers 0.20.3
albertus-sussex/veriscrape-simcse-book-reference_1_to_verify_9-fold-7
albertus-sussex
2025-03-25T18:25:03Z
0
0
transformers
[ "transformers", "safetensors", "roberta", "feature-extraction", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
feature-extraction
2025-03-25T18:24:37Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
LHRuig/damonebanksx
LHRuig
2025-03-25T18:24:47Z
0
0
diffusers
[ "diffusers", "safetensors", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T18:24:30Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: damonebanksx --- # damonebanksx <Gallery /> ## Model description damonebanksx lora ## Trigger words You should use `damonebanksx` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/damonebanksx/tree/main) them in the Files & versions tab.
HarshitSutrave/anomaly-detector-timesformer
HarshitSutrave
2025-03-25T18:24:42Z
0
0
transformers
[ "transformers", "safetensors", "timesformer", "video-classification", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
video-classification
2025-03-25T18:24:30Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
LHRuig/rpbrynsx
LHRuig
2025-03-25T18:24:09Z
0
0
diffusers
[ "diffusers", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T18:23:49Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: rpbrynsx --- # rpbrynsx <Gallery /> ## Model description rpbrynsx lora ## Trigger words You should use `rpbrynsx` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/rpbrynsx/tree/main) them in the Files & versions tab.
pictgensupport/birthdaycakev2
pictgensupport
2025-03-25T18:23:52Z
0
0
diffusers
[ "diffusers", "flux", "lora", "replicate", "text-to-image", "en", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "license:other", "region:us" ]
text-to-image
2025-03-25T18:23:49Z
--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md language: - en tags: - flux - diffusers - lora - replicate base_model: "black-forest-labs/FLUX.1-dev" pipeline_tag: text-to-image # widget: # - text: >- # prompt # output: # url: https://... instance_prompt: ICON_BASIC --- # Birthdaycakev2 <Gallery /> Trained on Replicate using: https://replicate.com/ostris/flux-dev-lora-trainer/train ## Trigger words You should use `ICON_BASIC` to trigger the image generation. ## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers) ```py from diffusers import AutoPipelineForText2Image import torch pipeline = AutoPipelineForText2Image.from_pretrained('black-forest-labs/FLUX.1-dev', torch_dtype=torch.float16).to('cuda') pipeline.load_lora_weights('pictgensupport/birthdaycakev2', weight_name='lora.safetensors') image = pipeline('your prompt').images[0] ``` For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters)
LHRuig/ainsleysx
LHRuig
2025-03-25T18:23:34Z
0
0
diffusers
[ "diffusers", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T18:23:15Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: ainsleysx --- # ainsleysx <Gallery /> ## Model description ainsleysx lora ## Trigger words You should use `ainsleysx` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/ainsleysx/tree/main) them in the Files & versions tab.
lesso01/c973deae-5754-433d-a9fe-fc2e168173ca
lesso01
2025-03-25T18:23:21Z
0
0
peft
[ "peft", "safetensors", "llama", "axolotl", "generated_from_trainer", "custom_code", "base_model:NousResearch/Yarn-Solar-10b-32k", "base_model:adapter:NousResearch/Yarn-Solar-10b-32k", "license:apache-2.0", "region:us" ]
null
2025-03-25T13:22:09Z
--- library_name: peft license: apache-2.0 base_model: NousResearch/Yarn-Solar-10b-32k tags: - axolotl - generated_from_trainer model-index: - name: c973deae-5754-433d-a9fe-fc2e168173ca results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.4.1` ```yaml adapter: lora base_model: NousResearch/Yarn-Solar-10b-32k bf16: auto chat_template: llama3 dataset_prepared_path: null datasets: - data_files: - 65d9e80afe69aff1_train_data.json ds_type: json format: custom path: /workspace/input_data/65d9e80afe69aff1_train_data.json type: field_input: documents field_instruction: question field_output: answer format: '{instruction} {input}' no_input_format: '{instruction}' system_format: '{system}' system_prompt: '' debug: null deepspeed: null do_eval: true early_stopping_patience: 3 eval_batch_size: 4 eval_max_new_tokens: 128 eval_steps: 500 evals_per_epoch: null flash_attention: true fp16: false fsdp: null fsdp_config: null gradient_accumulation_steps: 8 gradient_checkpointing: true group_by_length: true hub_model_id: lesso01/c973deae-5754-433d-a9fe-fc2e168173ca hub_repo: null hub_strategy: checkpoint hub_token: null learning_rate: 0.000201 load_in_4bit: false load_in_8bit: false local_rank: null logging_steps: 50 lora_alpha: 128 lora_dropout: 0.15 lora_fan_in_fan_out: null lora_model_dir: null lora_r: 64 lora_target_linear: true lr_scheduler: cosine max_grad_norm: 1.0 max_steps: 500 micro_batch_size: 4 mlflow_experiment_name: /tmp/65d9e80afe69aff1_train_data.json model_type: AutoModelForCausalLM num_epochs: 10 optimizer: adamw_torch_fused output_dir: miner_id_24 pad_to_sequence_len: true resume_from_checkpoint: null s2_attention: null sample_packing: false save_steps: 500 saves_per_epoch: null seed: 10 sequence_len: 1024 special_tokens: pad_token: </s> strict: false tf32: true tokenizer_type: AutoTokenizer train_on_inputs: false trust_remote_code: true val_set_size: 0.05 wandb_entity: null wandb_mode: online wandb_name: a1a050a4-6a01-49dd-9cd7-289119b180f3 wandb_project: 01a wandb_run: your_name wandb_runid: a1a050a4-6a01-49dd-9cd7-289119b180f3 warmup_steps: 100 weight_decay: 0.0 xformers_attention: null ``` </details><br> # c973deae-5754-433d-a9fe-fc2e168173ca This model is a fine-tuned version of [NousResearch/Yarn-Solar-10b-32k](https://huggingface.co/NousResearch/Yarn-Solar-10b-32k) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.0228 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.000201 - train_batch_size: 4 - eval_batch_size: 4 - seed: 10 - gradient_accumulation_steps: 8 - total_train_batch_size: 32 - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 100 - training_steps: 500 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:------:|:----:|:---------------:| | No log | 0.0008 | 1 | 1.6909 | | 8.1874 | 0.4238 | 500 | 1.0228 | ### Framework versions - PEFT 0.13.2 - Transformers 4.46.0 - Pytorch 2.5.0+cu124 - Datasets 3.0.1 - Tokenizers 0.20.1
albertus-sussex/veriscrape-simcse-book-reference_1_to_verify_9-fold-5
albertus-sussex
2025-03-25T18:23:21Z
0
0
transformers
[ "transformers", "safetensors", "roberta", "feature-extraction", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
feature-extraction
2025-03-25T18:22:56Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
SAPNA-SHAH-M-E/sapna.shah.viral.nxn.video.clip
SAPNA-SHAH-M-E
2025-03-25T18:22:45Z
0
0
null
[ "region:us" ]
null
2025-03-25T18:22:27Z
<animated-image data-catalyst=""><a href="https://alltvsteam.com/viral-video/?v=news-es-tvdf" rel="nofollow" data-target="animated-image.originalLink"><img src="https://static.wixstatic.com/media/b249f9_adac8f70fb3f45b88691696c77de18f3~mv2.gif" alt="Foo" data-canonical-src="https://static.wixstatic.com/media/b249f9_adac8f70fb3f45b88691696c77de18f3~mv2.gif" style="max-width: 100%; display: inline-block;" data-target="animated-image.originalImage"></a>
albertus-sussex/veriscrape-simcse-book-reference_1_to_verify_9-fold-4
albertus-sussex
2025-03-25T18:22:30Z
0
0
transformers
[ "transformers", "safetensors", "roberta", "feature-extraction", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
feature-extraction
2025-03-25T18:22:03Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
chukwuagoziesolomon/crypto-chatbot
chukwuagoziesolomon
2025-03-25T18:21:38Z
35
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:TinyLlama/TinyLlama-1.1B-Chat-v1.0", "base_model:adapter:TinyLlama/TinyLlama-1.1B-Chat-v1.0", "region:us" ]
null
2025-03-14T12:07:59Z
--- base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0 library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.14.0
arzuhussein/atllama.v3.5
arzuhussein
2025-03-25T18:19:51Z
0
0
null
[ "safetensors", "llama", "azerbaijani", "alpaca", "az", "base_model:meta-llama/Llama-3.1-8B-Instruct", "base_model:finetune:meta-llama/Llama-3.1-8B-Instruct", "license:mit", "region:us" ]
null
2024-12-26T16:01:50Z
--- license: mit language: - az base_model: - meta-llama/Llama-3.1-8B-Instruct tags: - azerbaijani - alpaca - az --- # Model Card for Atllama Atllama (Azerbaijani Tuned LLaMA) is a fine-tuned language model, specifically designed to improve instruction-following, comprehension, and text generation in the Azerbaijani language. It is part of an experimental project aimed at building a suite of Azerbaijani-focused NLP tools and models. This model card provides a comprehensive overview of Atllama, its development process, intended use cases, risks, and technical specifications. ## Model Details ### Model Description Atllama is an Azerbaijani fine-tuned version of the LLaMA model, developed as part of an experimental effort to enhance Azerbaijani language understanding and generation capabilities. The project explores ways to improve NLP tools in underrepresented languages like Azerbaijani, with Atllama being a core component for language-based applications. - **Developed by:** Arzu Huseynov and Nigar Arabli - **Funded by [optional]:** Self-funded - **Shared by [optional]:** Arzu Huseynov - **Model type:** Fine-tuned LLaMA (Azerbaijani) - **Language(s) (NLP):** Azerbaijani - **License:** Open-source, MIT - **Finetuned from model:** LLaMA 3.1 8B model ### Model Sources [optional] - **Repository:** [Add link when available] - **Paper [optional]:** [Add paper if available] - **Demo [optional]:** [Add demo link if available] ## GGUF Format Support Atllama is also available in the GGUF (GPT-Generated Unified Format) file format, which allows users to efficiently run the model on local machines using frameworks like `llama.cpp`, `Ollama`, or other GGML-based inference libraries. GGUF is an ideal format for lightweight inference, and the file includes both the model weights and metadata, enabling faster loading and usage with minimal setup. Users can find the GGUF files for Atllama in the repository, and here is how to run it: ### Example Usage with GGUF To run Atllama in the GGUF format on your local machine: 1. Download the GGUF file from the Hugging Face repository. 2. Use tools like `llama.cpp` or `Ollama` to load the model: ```bash ollama run atllama.gguf "Your Azerbaijani input prompt here" ``` For detailed instructions on GGUF and its usage with local inference tools, please refer to the respective documentation for `llama.cpp` and `Ollama` tools. ## Uses Atllama is designed to be used in various NLP tasks that require Azerbaijani language processing, including text generation, question-answer systems, instruction-following, and more. ### Direct Use Atllama can be directly used for: - Azerbaijani text generation - Following Azerbaijani-language instructions - Question-answer systems for Azerbaijani ### Downstream Use [optional] When fine-tuned further, Atllama can be adapted to: - Improve conversational agents for Azerbaijani-speaking users - Generate datasets specific to Azerbaijani NLP tasks - Assist in text correction or translation efforts in Azerbaijani ### Out-of-Scope Use The model may not perform well for: - Non-Azerbaijani language tasks - Domains where highly specific contextual knowledge is required (e.g., scientific data or legal texts outside of Azerbaijani context) ## Bias, Risks, and Limitations Atllama, like other fine-tuned models, may carry certain biases from the dataset it was trained on. These biases can affect: - Representation of minority groups or underrepresented topics in Azerbaijani contexts - Language model accuracy in specific dialects or regional variations of Azerbaijani ### Recommendations Users should be cautious of potential biases, particularly when using the model for sensitive content or high-stakes applications. More detailed testing across different subpopulations in Azerbaijani-speaking regions is recommended to mitigate risks. ## Training Details ### Training Data Atllama3.5 was trained using a variety of Azerbaijani text sources, including Wikipedia, news articles, and custom datasets. The training data was carefully curated to cover diverse topics, but there may still be limitations in niche domains. - **Dataset:** A 50K example dataset including instructional pairs and Wikipedia data. ### Training Procedure The model was fine-tuned using: - **Hardware:** PC (96GB RAM, RTX 4090, i9 CPU) - **Training regime:** fp16 mixed precision - **Epochs:** 3 epochs with additional fine-tuning for task-specific improvements #### Preprocessing Text data was cleaned for grammatical accuracy and translated from English sources in some cases, ensuring a focus on Azerbaijani language instruction-following. ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data Atllama was tested on custom datasets and Azerbaijani conversational tasks to evaluate its performance in instruction-following and text generation. #### Factors The model was evaluated across various factors, such as: - Comprehension of formal vs. colloquial Azerbaijani - Performance in generating coherent Azerbaijani instructions - Quality of output in terms of grammar and contextual relevance #### Metrics Evaluation metrics include: - Accuracy in instruction-following tasks - Fluency of generated text - User satisfaction in conversational contexts ### Results Atllama3.5 has shown significant improvement in understanding instructions and generating more accurate Azerbaijani text. However, the model may still struggle with edge cases involving regional dialects or very specific domains. Please keep in mind this model is not intended for production use in its current state. #### Summary Atllama3.5 continues to evolve as part of ongoing research into Azerbaijani language processing. While promising in its current form, future iterations aim to address biases and limitations. ## Environmental Impact - **Hardware Type:** Personal machine with high-end specs (96GB RAM, RTX 4090, i9 CPU) - **Hours used:** More than 100 hours+ - **Cloud Provider:** N/A (on-premises training) - **Compute Region:** N/A - **Carbon Emitted:** N/A ## Technical Specifications [optional] ### Model Architecture and Objective Atllama is based on LLaMA 3.1 architecture, fine-tuned for Azerbaijani NLP tasks with the objective of improving instruction-following and text generation. ### Compute Infrastructure The model was trained on a high-end local machine, as described in the "Training Procedure" section. ## Citation [optional] **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] - **LLaMA:** A family of language models designed by Meta, used as the base for fine-tuning in specific languages like Azerbaijani. - **Fine-tuning:** The process of adapting a pre-trained model to specific tasks or languages. ## More Information [optional] For more information, reach out to Arzu. ## Model Card Authors [optional] Arzu Huseynov [[email protected]], Nigar Arabli [[email protected]] ## Model Card Contact Feel free to reach out to me for collaboration or questions at [[email protected]].
albertus-sussex/veriscrape-simcse-book-reference_1_to_verify_9-fold-1
albertus-sussex
2025-03-25T18:19:48Z
0
0
transformers
[ "transformers", "safetensors", "roberta", "feature-extraction", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
feature-extraction
2025-03-25T18:19:14Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
Dataologist/gte_fine_tuned
Dataologist
2025-03-25T18:18:55Z
0
0
sentence-transformers
[ "sentence-transformers", "safetensors", "bert", "sentence-similarity", "feature-extraction", "generated_from_trainer", "dataset_size:10748192", "loss:MultipleNegativesRankingLoss", "loss:SoftmaxLoss", "arxiv:1908.10084", "arxiv:1705.00652", "autotrain_compatible", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
sentence-similarity
2025-03-25T18:17:45Z
--- tags: - sentence-transformers - sentence-similarity - feature-extraction - generated_from_trainer - dataset_size:10748192 - loss:MultipleNegativesRankingLoss - loss:SoftmaxLoss widget: - source_sentence: Burning Sensation sentences: - Diana disappeared 6 months ago. Now several people will join to solve this twisted story - James Tien Chun - Akira Takarada - source_sentence: Bagasi sentences: - Movie - T.V./Show - Bruce Willis - source_sentence: My Brother and I sentences: - Movie - Gastone Moschin - Sebastià Gasch - source_sentence: '10:16' sentences: - Movie - Matthew Gray Gubler - Isra Elsalihie - source_sentence: Cobra sentences: - Penny Pax - Maniyanpilla Raju - Movie pipeline_tag: sentence-similarity library_name: sentence-transformers --- # SentenceTransformer This is a [sentence-transformers](https://www.SBERT.net) model trained. It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more. ## Model Details ### Model Description - **Model Type:** Sentence Transformer <!-- - **Base model:** [Unknown](https://huggingface.co/unknown) --> - **Maximum Sequence Length:** 512 tokens - **Output Dimensionality:** 1024 dimensions - **Similarity Function:** Cosine Similarity <!-- - **Training Dataset:** Unknown --> <!-- - **Language:** Unknown --> <!-- - **License:** Unknown --> ### Model Sources - **Documentation:** [Sentence Transformers Documentation](https://sbert.net) - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers) - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers) ### Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel (1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True}) (2): Normalize() ) ``` ## Usage ### Direct Usage (Sentence Transformers) First install the Sentence Transformers library: ```bash pip install -U sentence-transformers ``` Then you can load this model and run inference. ```python from sentence_transformers import SentenceTransformer # Download from the 🤗 Hub model = SentenceTransformer("sentence_transformers_model_id") # Run inference sentences = [ 'Cobra', 'Maniyanpilla Raju', 'Movie', ] embeddings = model.encode(sentences) print(embeddings.shape) # [3, 1024] # Get the similarity scores for the embeddings similarities = model.similarity(embeddings, embeddings) print(similarities.shape) # [3, 3] ``` <!-- ### Direct Usage (Transformers) <details><summary>Click to see the direct usage in Transformers</summary> </details> --> <!-- ### Downstream Usage (Sentence Transformers) You can finetune this model on your own dataset. <details><summary>Click to expand</summary> </details> --> <!-- ### Out-of-Scope Use *List how the model may foreseeably be misused and address what users ought not to do with the model.* --> <!-- ## Bias, Risks and Limitations *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.* --> <!-- ### Recommendations *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.* --> ## Training Details ### Training Datasets #### Unnamed Dataset * Size: 5,374,096 training samples * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>label</code> * Approximate statistics based on the first 1000 samples: | | sentence_0 | sentence_1 | label | |:--------|:---------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:------------------------------------------------| | type | string | string | int | | details | <ul><li>min: 3 tokens</li><li>mean: 6.61 tokens</li><li>max: 41 tokens</li></ul> | <ul><li>min: 3 tokens</li><li>mean: 14.0 tokens</li><li>max: 212 tokens</li></ul> | <ul><li>0: ~20.70%</li><li>1: ~79.30%</li></ul> | * Samples: | sentence_0 | sentence_1 | label | |:-----------------------------|:-------------------------------|:---------------| | <code>Film Fun</code> | <code>Movie</code> | <code>1</code> | | <code>Fighter</code> | <code>Anna Karczmarczyk</code> | <code>1</code> | | <code>The Water Nymph</code> | <code>rating: 4.5</code> | <code>1</code> | * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: ```json { "scale": 20.0, "similarity_fct": "cos_sim" } ``` #### Unnamed Dataset * Size: 5,374,096 training samples * Columns: <code>sentence_0</code>, <code>sentence_1</code>, and <code>label</code> * Approximate statistics based on the first 1000 samples: | | sentence_0 | sentence_1 | label | |:--------|:---------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:------------------------------------------------| | type | string | string | int | | details | <ul><li>min: 3 tokens</li><li>mean: 6.53 tokens</li><li>max: 24 tokens</li></ul> | <ul><li>min: 3 tokens</li><li>mean: 13.94 tokens</li><li>max: 236 tokens</li></ul> | <ul><li>0: ~19.90%</li><li>1: ~80.10%</li></ul> | * Samples: | sentence_0 | sentence_1 | label | |:---------------------------------------------------------|:----------------------------------|:---------------| | <code>Humans</code> | <code>Dominique Pinon</code> | <code>1</code> | | <code>Two Strangers Trying Not to Kill Each Other</code> | <code>Billy Chow Bei-Lei</code> | <code>0</code> | | <code>Single Ladies</code> | <code>Harold 'House' Moore</code> | <code>1</code> | * Loss: [<code>SoftmaxLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#softmaxloss) ### Training Hyperparameters #### Non-Default Hyperparameters - `per_device_train_batch_size`: 16 - `per_device_eval_batch_size`: 16 - `num_train_epochs`: 0.1 - `multi_dataset_batch_sampler`: round_robin #### All Hyperparameters <details><summary>Click to expand</summary> - `overwrite_output_dir`: False - `do_predict`: False - `eval_strategy`: no - `prediction_loss_only`: True - `per_device_train_batch_size`: 16 - `per_device_eval_batch_size`: 16 - `per_gpu_train_batch_size`: None - `per_gpu_eval_batch_size`: None - `gradient_accumulation_steps`: 1 - `eval_accumulation_steps`: None - `torch_empty_cache_steps`: None - `learning_rate`: 5e-05 - `weight_decay`: 0.0 - `adam_beta1`: 0.9 - `adam_beta2`: 0.999 - `adam_epsilon`: 1e-08 - `max_grad_norm`: 1 - `num_train_epochs`: 0.1 - `max_steps`: -1 - `lr_scheduler_type`: linear - `lr_scheduler_kwargs`: {} - `warmup_ratio`: 0.0 - `warmup_steps`: 0 - `log_level`: passive - `log_level_replica`: warning - `log_on_each_node`: True - `logging_nan_inf_filter`: True - `save_safetensors`: True - `save_on_each_node`: False - `save_only_model`: False - `restore_callback_states_from_checkpoint`: False - `no_cuda`: False - `use_cpu`: False - `use_mps_device`: False - `seed`: 42 - `data_seed`: None - `jit_mode_eval`: False - `use_ipex`: False - `bf16`: False - `fp16`: False - `fp16_opt_level`: O1 - `half_precision_backend`: auto - `bf16_full_eval`: False - `fp16_full_eval`: False - `tf32`: None - `local_rank`: 0 - `ddp_backend`: None - `tpu_num_cores`: None - `tpu_metrics_debug`: False - `debug`: [] - `dataloader_drop_last`: False - `dataloader_num_workers`: 0 - `dataloader_prefetch_factor`: None - `past_index`: -1 - `disable_tqdm`: False - `remove_unused_columns`: True - `label_names`: None - `load_best_model_at_end`: False - `ignore_data_skip`: False - `fsdp`: [] - `fsdp_min_num_params`: 0 - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False} - `fsdp_transformer_layer_cls_to_wrap`: None - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None} - `deepspeed`: None - `label_smoothing_factor`: 0.0 - `optim`: adamw_torch - `optim_args`: None - `adafactor`: False - `group_by_length`: False - `length_column_name`: length - `ddp_find_unused_parameters`: None - `ddp_bucket_cap_mb`: None - `ddp_broadcast_buffers`: False - `dataloader_pin_memory`: True - `dataloader_persistent_workers`: False - `skip_memory_metrics`: True - `use_legacy_prediction_loop`: False - `push_to_hub`: False - `resume_from_checkpoint`: None - `hub_model_id`: None - `hub_strategy`: every_save - `hub_private_repo`: None - `hub_always_push`: False - `gradient_checkpointing`: False - `gradient_checkpointing_kwargs`: None - `include_inputs_for_metrics`: False - `include_for_metrics`: [] - `eval_do_concat_batches`: True - `fp16_backend`: auto - `push_to_hub_model_id`: None - `push_to_hub_organization`: None - `mp_parameters`: - `auto_find_batch_size`: False - `full_determinism`: False - `torchdynamo`: None - `ray_scope`: last - `ddp_timeout`: 1800 - `torch_compile`: False - `torch_compile_backend`: None - `torch_compile_mode`: None - `dispatch_batches`: None - `split_batches`: None - `include_tokens_per_second`: False - `include_num_input_tokens_seen`: False - `neftune_noise_alpha`: None - `optim_target_modules`: None - `batch_eval_metrics`: False - `eval_on_start`: False - `use_liger_kernel`: False - `eval_use_gather_object`: False - `average_tokens_across_devices`: False - `prompts`: None - `batch_sampler`: batch_sampler - `multi_dataset_batch_sampler`: round_robin </details> ### Training Logs <details><summary>Click to expand</summary> | Epoch | Step | Training Loss | |:------:|:-----:|:-------------:| | 0.0007 | 500 | 1.6186 | | 0.0015 | 1000 | 1.6232 | | 0.0022 | 1500 | 1.6157 | | 0.0030 | 2000 | 1.6164 | | 0.0037 | 2500 | 1.6139 | | 0.0045 | 3000 | 1.6133 | | 0.0052 | 3500 | 1.622 | | 0.0060 | 4000 | 1.6258 | | 0.0067 | 4500 | 1.6091 | | 0.0074 | 5000 | 1.6141 | | 0.0082 | 5500 | 1.6265 | | 0.0089 | 6000 | 1.6048 | | 0.0097 | 6500 | 1.6185 | | 0.0104 | 7000 | 1.6053 | | 0.0112 | 7500 | 1.6126 | | 0.0119 | 8000 | 1.6198 | | 0.0127 | 8500 | 1.6076 | | 0.0134 | 9000 | 1.6139 | | 0.0141 | 9500 | 1.6112 | | 0.0149 | 10000 | 1.6215 | | 0.0156 | 10500 | 1.611 | | 0.0164 | 11000 | 1.6074 | | 0.0171 | 11500 | 1.6232 | | 0.0179 | 12000 | 1.6128 | | 0.0186 | 12500 | 1.6145 | | 0.0194 | 13000 | 1.6133 | | 0.0201 | 13500 | 1.6129 | | 0.0208 | 14000 | 1.6197 | | 0.0216 | 14500 | 1.6071 | | 0.0223 | 15000 | 1.6066 | | 0.0231 | 15500 | 1.6126 | | 0.0238 | 16000 | 1.6115 | | 0.0246 | 16500 | 1.6179 | | 0.0253 | 17000 | 1.6112 | | 0.0261 | 17500 | 1.609 | | 0.0268 | 18000 | 1.6148 | | 0.0275 | 18500 | 1.6094 | | 0.0283 | 19000 | 1.6107 | | 0.0290 | 19500 | 1.6075 | | 0.0298 | 20000 | 1.6064 | | 0.0305 | 20500 | 1.612 | | 0.0313 | 21000 | 1.62 | | 0.0320 | 21500 | 1.6006 | | 0.0327 | 22000 | 1.6193 | | 0.0335 | 22500 | 1.6042 | | 0.0342 | 23000 | 1.6076 | | 0.0350 | 23500 | 1.6083 | | 0.0357 | 24000 | 1.6015 | | 0.0365 | 24500 | 1.6167 | | 0.0372 | 25000 | 1.6141 | | 0.0380 | 25500 | 1.6115 | | 0.0387 | 26000 | 1.6176 | | 0.0394 | 26500 | 1.6054 | | 0.0402 | 27000 | 1.5942 | | 0.0409 | 27500 | 1.6067 | | 0.0417 | 28000 | 1.6079 | | 0.0424 | 28500 | 1.6013 | | 0.0432 | 29000 | 1.6063 | | 0.0439 | 29500 | 1.6079 | | 0.0447 | 30000 | 1.6144 | | 0.0454 | 30500 | 1.5996 | | 0.0461 | 31000 | 1.6088 | | 0.0469 | 31500 | 1.6058 | | 0.0476 | 32000 | 1.6109 | | 0.0484 | 32500 | 1.6081 | | 0.0491 | 33000 | 1.5996 | | 0.0499 | 33500 | 1.6031 | | 0.0506 | 34000 | 1.6146 | | 0.0514 | 34500 | 1.6035 | | 0.0521 | 35000 | 1.6086 | | 0.0528 | 35500 | 1.6144 | | 0.0536 | 36000 | 1.6081 | | 0.0543 | 36500 | 1.6062 | | 0.0551 | 37000 | 1.6126 | | 0.0558 | 37500 | 1.604 | | 0.0566 | 38000 | 1.6026 | | 0.0573 | 38500 | 1.6093 | | 0.0581 | 39000 | 1.606 | | 0.0588 | 39500 | 1.6143 | | 0.0595 | 40000 | 1.6017 | | 0.0603 | 40500 | 1.6174 | | 0.0610 | 41000 | 1.6166 | | 0.0618 | 41500 | 1.6086 | | 0.0625 | 42000 | 1.6125 | | 0.0633 | 42500 | 1.6102 | | 0.0640 | 43000 | 1.6047 | | 0.0648 | 43500 | 1.6067 | | 0.0655 | 44000 | 1.612 | | 0.0662 | 44500 | 1.6064 | | 0.0670 | 45000 | 1.6069 | | 0.0677 | 45500 | 1.6071 | | 0.0685 | 46000 | 1.6012 | | 0.0692 | 46500 | 1.6034 | | 0.0700 | 47000 | 1.6224 | | 0.0707 | 47500 | 1.6115 | | 0.0715 | 48000 | 1.6091 | | 0.0722 | 48500 | 1.6063 | | 0.0729 | 49000 | 1.6105 | | 0.0737 | 49500 | 1.5979 | | 0.0744 | 50000 | 1.6175 | | 0.0752 | 50500 | 1.6066 | | 0.0759 | 51000 | 1.6114 | | 0.0767 | 51500 | 1.6096 | | 0.0774 | 52000 | 1.6078 | | 0.0782 | 52500 | 1.6008 | | 0.0789 | 53000 | 1.6075 | | 0.0796 | 53500 | 1.6069 | | 0.0804 | 54000 | 1.6088 | | 0.0811 | 54500 | 1.6076 | | 0.0819 | 55000 | 1.6047 | | 0.0826 | 55500 | 1.6087 | | 0.0834 | 56000 | 1.6202 | | 0.0841 | 56500 | 1.6052 | | 0.0849 | 57000 | 1.6123 | | 0.0856 | 57500 | 1.5969 | | 0.0863 | 58000 | 1.6053 | | 0.0871 | 58500 | 1.6096 | | 0.0878 | 59000 | 1.6083 | | 0.0886 | 59500 | 1.6018 | | 0.0893 | 60000 | 1.6066 | | 0.0901 | 60500 | 1.6187 | | 0.0908 | 61000 | 1.604 | | 0.0916 | 61500 | 1.6041 | | 0.0923 | 62000 | 1.608 | | 0.0930 | 62500 | 1.602 | | 0.0938 | 63000 | 1.6003 | | 0.0945 | 63500 | 1.614 | | 0.0953 | 64000 | 1.6162 | | 0.0960 | 64500 | 1.6056 | | 0.0968 | 65000 | 1.6124 | | 0.0975 | 65500 | 1.6203 | | 0.0982 | 66000 | 1.6092 | | 0.0990 | 66500 | 1.6027 | | 0.0997 | 67000 | 1.606 | | 0.0007 | 500 | 1.6089 | | 0.0015 | 1000 | 1.6112 | | 0.0022 | 1500 | 1.6097 | | 0.0030 | 2000 | 1.5993 | | 0.0037 | 2500 | 1.6027 | | 0.0045 | 3000 | 1.6081 | | 0.0052 | 3500 | 1.6057 | | 0.0060 | 4000 | 1.6168 | | 0.0067 | 4500 | 1.603 | | 0.0074 | 5000 | 1.6025 | | 0.0082 | 5500 | 1.6049 | | 0.0089 | 6000 | 1.6066 | | 0.0097 | 6500 | 1.6049 | | 0.0104 | 7000 | 1.6004 | | 0.0112 | 7500 | 1.6038 | | 0.0119 | 8000 | 1.6015 | | 0.0127 | 8500 | 1.6081 | | 0.0134 | 9000 | 1.6075 | | 0.0141 | 9500 | 1.5987 | | 0.0149 | 10000 | 1.6061 | | 0.0156 | 10500 | 1.599 | | 0.0164 | 11000 | 1.6107 | | 0.0171 | 11500 | 1.6144 | | 0.0179 | 12000 | 1.6058 | | 0.0186 | 12500 | 1.6062 | | 0.0194 | 13000 | 1.6015 | | 0.0201 | 13500 | 1.6006 | | 0.0208 | 14000 | 1.6058 | | 0.0216 | 14500 | 1.6063 | | 0.0223 | 15000 | 1.5987 | | 0.0231 | 15500 | 1.6059 | | 0.0238 | 16000 | 1.6068 | | 0.0246 | 16500 | 1.6013 | | 0.0253 | 17000 | 1.6013 | | 0.0261 | 17500 | 1.5933 | | 0.0268 | 18000 | 1.6066 | | 0.0275 | 18500 | 1.6042 | | 0.0283 | 19000 | 1.5953 | | 0.0290 | 19500 | 1.5999 | | 0.0298 | 20000 | 1.6084 | | 0.0305 | 20500 | 1.5982 | | 0.0313 | 21000 | 1.6016 | | 0.0320 | 21500 | 1.6047 | | 0.0327 | 22000 | 1.6036 | | 0.0335 | 22500 | 1.5971 | | 0.0342 | 23000 | 1.6055 | | 0.0350 | 23500 | 1.6081 | | 0.0357 | 24000 | 1.6005 | | 0.0365 | 24500 | 1.6031 | | 0.0372 | 25000 | 1.5923 | | 0.0380 | 25500 | 1.604 | | 0.0387 | 26000 | 1.6057 | | 0.0394 | 26500 | 1.6001 | | 0.0402 | 27000 | 1.6016 | | 0.0409 | 27500 | 1.6073 | | 0.0417 | 28000 | 1.6071 | | 0.0424 | 28500 | 1.5928 | | 0.0432 | 29000 | 1.5985 | | 0.0439 | 29500 | 1.5915 | | 0.0447 | 30000 | 1.5937 | | 0.0454 | 30500 | 1.6056 | | 0.0461 | 31000 | 1.5975 | | 0.0469 | 31500 | 1.6036 | | 0.0476 | 32000 | 1.6043 | | 0.0484 | 32500 | 1.5967 | | 0.0491 | 33000 | 1.5973 | | 0.0499 | 33500 | 1.5899 | | 0.0506 | 34000 | 1.607 | | 0.0514 | 34500 | 1.5988 | | 0.0521 | 35000 | 1.5957 | | 0.0528 | 35500 | 1.6038 | | 0.0536 | 36000 | 1.5964 | | 0.0543 | 36500 | 1.6008 | | 0.0551 | 37000 | 1.6017 | | 0.0558 | 37500 | 1.6082 | | 0.0566 | 38000 | 1.5956 | | 0.0573 | 38500 | 1.5914 | | 0.0581 | 39000 | 1.5949 | | 0.0588 | 39500 | 1.5993 | | 0.0595 | 40000 | 1.6002 | | 0.0603 | 40500 | 1.5914 | | 0.0610 | 41000 | 1.5958 | | 0.0618 | 41500 | 1.6029 | | 0.0625 | 42000 | 1.6021 | | 0.0633 | 42500 | 1.5987 | | 0.0640 | 43000 | 1.5962 | | 0.0648 | 43500 | 1.5922 | | 0.0655 | 44000 | 1.6015 | | 0.0662 | 44500 | 1.5997 | | 0.0670 | 45000 | 1.596 | | 0.0677 | 45500 | 1.605 | | 0.0685 | 46000 | 1.5991 | | 0.0692 | 46500 | 1.5993 | | 0.0700 | 47000 | 1.5987 | | 0.0707 | 47500 | 1.6062 | | 0.0715 | 48000 | 1.5982 | | 0.0722 | 48500 | 1.6023 | | 0.0729 | 49000 | 1.6086 | | 0.0737 | 49500 | 1.5913 | | 0.0744 | 50000 | 1.5965 | | 0.0752 | 50500 | 1.6015 | | 0.0759 | 51000 | 1.598 | | 0.0767 | 51500 | 1.6034 | | 0.0774 | 52000 | 1.6089 | | 0.0782 | 52500 | 1.5924 | | 0.0789 | 53000 | 1.5959 | | 0.0796 | 53500 | 1.6045 | | 0.0804 | 54000 | 1.6011 | | 0.0811 | 54500 | 1.6048 | | 0.0819 | 55000 | 1.6052 | | 0.0826 | 55500 | 1.607 | | 0.0834 | 56000 | 1.5974 | | 0.0841 | 56500 | 1.5966 | | 0.0849 | 57000 | 1.5971 | | 0.0856 | 57500 | 1.6034 | | 0.0863 | 58000 | 1.599 | | 0.0871 | 58500 | 1.5975 | | 0.0878 | 59000 | 1.6017 | | 0.0886 | 59500 | 1.5985 | | 0.0893 | 60000 | 1.5984 | | 0.0901 | 60500 | 1.5934 | | 0.0908 | 61000 | 1.6042 | | 0.0916 | 61500 | 1.6032 | | 0.0923 | 62000 | 1.5972 | | 0.0930 | 62500 | 1.6005 | | 0.0938 | 63000 | 1.5987 | | 0.0945 | 63500 | 1.6036 | | 0.0953 | 64000 | 1.5944 | | 0.0960 | 64500 | 1.598 | | 0.0968 | 65000 | 1.6073 | | 0.0975 | 65500 | 1.6072 | | 0.0982 | 66000 | 1.5957 | | 0.0990 | 66500 | 1.603 | | 0.0997 | 67000 | 1.5908 | | 0.0007 | 500 | 1.57 | | 0.0015 | 1000 | 1.5611 | | 0.0022 | 1500 | 1.558 | | 0.0030 | 2000 | 1.5431 | | 0.0037 | 2500 | 1.5478 | | 0.0045 | 3000 | 1.553 | | 0.0052 | 3500 | 1.5507 | | 0.0060 | 4000 | 1.5631 | | 0.0067 | 4500 | 1.5513 | | 0.0074 | 5000 | 1.5504 | | 0.0082 | 5500 | 1.5552 | | 0.0089 | 6000 | 1.5564 | | 0.0097 | 6500 | 1.5574 | | 0.0104 | 7000 | 1.5522 | | 0.0112 | 7500 | 1.5577 | | 0.0119 | 8000 | 1.5559 | | 0.0127 | 8500 | 1.5623 | | 0.0134 | 9000 | 1.5629 | | 0.0141 | 9500 | 1.5547 | | 0.0149 | 10000 | 1.5639 | | 0.0156 | 10500 | 1.5567 | | 0.0164 | 11000 | 1.57 | | 0.0171 | 11500 | 1.5745 | | 0.0179 | 12000 | 1.5658 | | 0.0186 | 12500 | 1.5669 | | 0.0194 | 13000 | 1.5625 | | 0.0201 | 13500 | 1.5612 | | 0.0208 | 14000 | 1.5698 | | 0.0216 | 14500 | 1.5704 | | 0.0223 | 15000 | 1.5628 | | 0.0231 | 15500 | 1.5723 | | 0.0238 | 16000 | 1.5736 | | 0.0246 | 16500 | 1.5674 | | 0.0253 | 17000 | 1.5686 | | 0.0261 | 17500 | 1.5635 | | 0.0268 | 18000 | 1.5774 | | 0.0275 | 18500 | 1.5746 | | 0.0283 | 19000 | 1.5649 | | 0.0290 | 19500 | 1.5714 | | 0.0298 | 20000 | 1.5806 | | 0.0305 | 20500 | 1.5722 | | 0.0313 | 21000 | 1.5771 | | 0.0320 | 21500 | 1.5802 | | 0.0327 | 22000 | 1.5819 | | 0.0335 | 22500 | 1.5752 | | 0.0342 | 23000 | 1.5823 | | 0.0350 | 23500 | 1.5863 | | 0.0357 | 24000 | 1.5799 | | 0.0365 | 24500 | 1.5844 | | 0.0372 | 25000 | 1.5734 | | 0.0380 | 25500 | 1.5862 | | 0.0387 | 26000 | 1.5899 | | 0.0394 | 26500 | 1.584 | | 0.0402 | 27000 | 1.5875 | | 0.0409 | 27500 | 1.5914 | | 0.0417 | 28000 | 1.5921 | | 0.0424 | 28500 | 1.5812 | | 0.0432 | 29000 | 1.586 | | 0.0439 | 29500 | 1.5799 | | 0.0447 | 30000 | 1.5827 | | 0.0454 | 30500 | 1.5966 | | 0.0461 | 31000 | 1.5881 | | 0.0469 | 31500 | 1.5963 | | 0.0476 | 32000 | 1.5982 | | 0.0484 | 32500 | 1.5893 | | 0.0491 | 33000 | 1.5907 | | 0.0499 | 33500 | 1.5846 | | 0.0506 | 34000 | 1.6025 | | 0.0514 | 34500 | 1.5944 | | 0.0521 | 35000 | 1.591 | | 0.0528 | 35500 | 1.6 | | 0.0536 | 36000 | 1.5924 | | 0.0543 | 36500 | 1.5969 | | 0.0551 | 37000 | 1.5974 | | 0.0558 | 37500 | 1.6039 | | 0.0566 | 38000 | 1.5904 | | 0.0573 | 38500 | 1.5884 | | 0.0581 | 39000 | 1.5905 | | 0.0588 | 39500 | 1.5965 | | 0.0595 | 40000 | 1.5959 | | 0.0603 | 40500 | 1.5883 | | 0.0610 | 41000 | 1.5909 | | 0.0618 | 41500 | 1.5987 | | 0.0625 | 42000 | 1.5982 | | 0.0633 | 42500 | 1.5945 | | 0.0640 | 43000 | 1.592 | | 0.0648 | 43500 | 1.5886 | | 0.0655 | 44000 | 1.5974 | | 0.0662 | 44500 | 1.5964 | | 0.0670 | 45000 | 1.5907 | | 0.0677 | 45500 | 1.6007 | | 0.0685 | 46000 | 1.5935 | | 0.0692 | 46500 | 1.5949 | | 0.0700 | 47000 | 1.5945 | | 0.0707 | 47500 | 1.6033 | | 0.0715 | 48000 | 1.5935 | | 0.0722 | 48500 | 1.5982 | | 0.0729 | 49000 | 1.6039 | | 0.0737 | 49500 | 1.5861 | | 0.0744 | 50000 | 1.5924 | | 0.0752 | 50500 | 1.5966 | | 0.0759 | 51000 | 1.5952 | | 0.0767 | 51500 | 1.5992 | | 0.0774 | 52000 | 1.6043 | | 0.0782 | 52500 | 1.5876 | | 0.0789 | 53000 | 1.5912 | | 0.0796 | 53500 | 1.6013 | | 0.0804 | 54000 | 1.5979 | | 0.0811 | 54500 | 1.6016 | | 0.0819 | 55000 | 1.6013 | | 0.0826 | 55500 | 1.6031 | | 0.0834 | 56000 | 1.5935 | | 0.0841 | 56500 | 1.5923 | | 0.0849 | 57000 | 1.5918 | | 0.0856 | 57500 | 1.5992 | | 0.0863 | 58000 | 1.5949 | | 0.0871 | 58500 | 1.5947 | | 0.0878 | 59000 | 1.5973 | | 0.0886 | 59500 | 1.5935 | | 0.0893 | 60000 | 1.5947 | | 0.0901 | 60500 | 1.589 | | 0.0908 | 61000 | 1.6005 | | 0.0916 | 61500 | 1.598 | | 0.0923 | 62000 | 1.5937 | | 0.0930 | 62500 | 1.5965 | | 0.0938 | 63000 | 1.5953 | | 0.0945 | 63500 | 1.5992 | | 0.0953 | 64000 | 1.5892 | | 0.0960 | 64500 | 1.5946 | | 0.0968 | 65000 | 1.6038 | | 0.0975 | 65500 | 1.6038 | | 0.0982 | 66000 | 1.592 | | 0.0990 | 66500 | 1.5992 | | 0.0997 | 67000 | 1.5864 | | 0.0007 | 500 | 1.5298 | | 0.0015 | 1000 | 1.5082 | | 0.0022 | 1500 | 1.5018 | | 0.0030 | 2000 | 1.4843 | | 0.0037 | 2500 | 1.4893 | | 0.0045 | 3000 | 1.4945 | | 0.0052 | 3500 | 1.4915 | | 0.0060 | 4000 | 1.5063 | | 0.0067 | 4500 | 1.4954 | | 0.0074 | 5000 | 1.4951 | | 0.0082 | 5500 | 1.5004 | | 0.0089 | 6000 | 1.5014 | | 0.0097 | 6500 | 1.505 | | 0.0104 | 7000 | 1.4997 | | 0.0112 | 7500 | 1.5075 | | 0.0119 | 8000 | 1.5062 | | 0.0127 | 8500 | 1.5122 | | 0.0134 | 9000 | 1.5148 | | 0.0141 | 9500 | 1.5062 | | 0.0149 | 10000 | 1.5165 | | 0.0156 | 10500 | 1.5105 | | 0.0164 | 11000 | 1.5258 | | 0.0171 | 11500 | 1.5305 | | 0.0179 | 12000 | 1.5227 | | 0.0186 | 12500 | 1.5251 | | 0.0194 | 13000 | 1.5204 | | 0.0201 | 13500 | 1.5204 | | 0.0208 | 14000 | 1.5303 | | 0.0216 | 14500 | 1.5324 | | 0.0223 | 15000 | 1.524 | | 0.0231 | 15500 | 1.5358 | | 0.0238 | 16000 | 1.5372 | | 0.0246 | 16500 | 1.5322 | | 0.0253 | 17000 | 1.5346 | | 0.0261 | 17500 | 1.5328 | | 0.0268 | 18000 | 1.5474 | | 0.0275 | 18500 | 1.5436 | | 0.0283 | 19000 | 1.5339 | | 0.0290 | 19500 | 1.5416 | | 0.0298 | 20000 | 1.5525 | | 0.0305 | 20500 | 1.5455 | | 0.0313 | 21000 | 1.5518 | | 0.0320 | 21500 | 1.5559 | | 0.0327 | 22000 | 1.5596 | | 0.0335 | 22500 | 1.5523 | | 0.0342 | 23000 | 1.5594 | | 0.0350 | 23500 | 1.5641 | | 0.0357 | 24000 | 1.5589 | | 0.0365 | 24500 | 1.5658 | | 0.0372 | 25000 | 1.5545 | | 0.0380 | 25500 | 1.569 | | 0.0387 | 26000 | 1.5742 | | 0.0394 | 26500 | 1.5687 | | 0.0402 | 27000 | 1.5735 | | 0.0409 | 27500 | 1.5768 | | 0.0417 | 28000 | 1.5785 | | 0.0424 | 28500 | 1.5712 | | 0.0432 | 29000 | 1.5741 | | 0.0439 | 29500 | 1.5705 | | 0.0447 | 30000 | 1.5728 | | 0.0454 | 30500 | 1.5887 | | 0.0461 | 31000 | 1.5799 | | 0.0469 | 31500 | 1.5909 | | 0.0476 | 32000 | 1.5941 | | 0.0484 | 32500 | 1.5843 | | 0.0491 | 33000 | 1.5869 | | 0.0499 | 33500 | 1.582 | | 0.0506 | 34000 | 1.601 | | 0.0514 | 34500 | 1.5931 | | 0.0521 | 35000 | 1.5887 | | 0.0528 | 35500 | 1.5987 | | 0.0536 | 36000 | 1.5903 | | 0.0543 | 36500 | 1.5957 | | 0.0551 | 37000 | 1.5958 | | 0.0558 | 37500 | 1.6024 | | 0.0566 | 38000 | 1.5882 | | 0.0573 | 38500 | 1.5868 | | 0.0581 | 39000 | 1.5883 | | 0.0588 | 39500 | 1.5952 | | 0.0595 | 40000 | 1.594 | | 0.0603 | 40500 | 1.5864 | | 0.0610 | 41000 | 1.5881 | | 0.0618 | 41500 | 1.5966 | | 0.0625 | 42000 | 1.5966 | | 0.0633 | 42500 | 1.5934 | | 0.0640 | 43000 | 1.5906 | | 0.0648 | 43500 | 1.5867 | | 0.0655 | 44000 | 1.5959 | | 0.0662 | 44500 | 1.5955 | | 0.0670 | 45000 | 1.5886 | | 0.0677 | 45500 | 1.598 | | 0.0685 | 46000 | 1.5913 | | 0.0692 | 46500 | 1.5937 | | 0.0700 | 47000 | 1.593 | | 0.0707 | 47500 | 1.6021 | | 0.0715 | 48000 | 1.5907 | | 0.0722 | 48500 | 1.5977 | | 0.0729 | 49000 | 1.6012 | | 0.0737 | 49500 | 1.5838 | | 0.0744 | 50000 | 1.5912 | | 0.0752 | 50500 | 1.5942 | | 0.0759 | 51000 | 1.5941 | | 0.0767 | 51500 | 1.5972 | | 0.0774 | 52000 | 1.6029 | | 0.0782 | 52500 | 1.5851 | | 0.0789 | 53000 | 1.5891 | | 0.0796 | 53500 | 1.6 | | 0.0804 | 54000 | 1.5967 | | 0.0811 | 54500 | 1.6011 | | 0.0819 | 55000 | 1.6001 | | 0.0826 | 55500 | 1.6019 | | 0.0834 | 56000 | 1.5926 | | 0.0841 | 56500 | 1.5907 | | 0.0849 | 57000 | 1.5896 | | 0.0856 | 57500 | 1.5979 | | 0.0863 | 58000 | 1.5933 | | 0.0871 | 58500 | 1.594 | | 0.0878 | 59000 | 1.5958 | | 0.0886 | 59500 | 1.5913 | | 0.0893 | 60000 | 1.5938 | | 0.0901 | 60500 | 1.5872 | | 0.0908 | 61000 | 1.5992 | | 0.0916 | 61500 | 1.5955 | | 0.0923 | 62000 | 1.5926 | | 0.0930 | 62500 | 1.5948 | | 0.0938 | 63000 | 1.5936 | | 0.0945 | 63500 | 1.598 | | 0.0953 | 64000 | 1.5866 | | 0.0960 | 64500 | 1.5938 | | 0.0968 | 65000 | 1.6025 | | 0.0975 | 65500 | 1.6031 | | 0.0982 | 66000 | 1.5905 | | 0.0990 | 66500 | 1.5982 | | 0.0997 | 67000 | 1.5849 | | 0.0007 | 500 | 1.4875 | | 0.0015 | 1000 | 1.4515 | | 0.0022 | 1500 | 1.4396 | | 0.0030 | 2000 | 1.4207 | | 0.0037 | 2500 | 1.4276 | | 0.0045 | 3000 | 1.4314 | | 0.0052 | 3500 | 1.4281 | | 0.0060 | 4000 | 1.4451 | | 0.0067 | 4500 | 1.4352 | | 0.0074 | 5000 | 1.4358 | | 0.0082 | 5500 | 1.4409 | | 0.0089 | 6000 | 1.4418 | | 0.0097 | 6500 | 1.4475 | | 0.0104 | 7000 | 1.4435 | | 0.0112 | 7500 | 1.4512 | | 0.0119 | 8000 | 1.4519 | | 0.0127 | 8500 | 1.4573 | | 0.0134 | 9000 | 1.4619 | | 0.0141 | 9500 | 1.4529 | | 0.0149 | 10000 | 1.4655 | | 0.0156 | 10500 | 1.4608 | | 0.0164 | 11000 | 1.4782 | | 0.0171 | 11500 | 1.4827 | | 0.0179 | 12000 | 1.4776 | | 0.0186 | 12500 | 1.4808 | | 0.0194 | 13000 | 1.475 | | 0.0201 | 13500 | 1.4768 | | 0.0208 | 14000 | 1.4867 | | 0.0216 | 14500 | 1.4926 | | 0.0223 | 15000 | 1.4826 | | 0.0231 | 15500 | 1.4957 | | 0.0238 | 16000 | 1.4982 | | 0.0246 | 16500 | 1.4946 | | 0.0253 | 17000 | 1.4991 | | 0.0261 | 17500 | 1.4997 | | 0.0268 | 18000 | 1.5159 | | 0.0275 | 18500 | 1.5104 | | 0.0283 | 19000 | 1.5023 | | 0.0290 | 19500 | 1.5108 | | 0.0298 | 20000 | 1.5229 | | 0.0305 | 20500 | 1.5178 | | 0.0313 | 21000 | 1.5258 | | 0.0320 | 21500 | 1.5307 | | 0.0327 | 22000 | 1.5363 | | 0.0335 | 22500 | 1.5276 | | 0.0342 | 23000 | 1.5364 | | 0.0350 | 23500 | 1.5424 | | 0.0357 | 24000 | 1.5382 | | 0.0365 | 24500 | 1.547 | | 0.0372 | 25000 | 1.5357 | | 0.0380 | 25500 | 1.5521 | | 0.0387 | 26000 | 1.5594 | | 0.0394 | 26500 | 1.5539 | | 0.0402 | 27000 | 1.5599 | | 0.0409 | 27500 | 1.5626 | | 0.0417 | 28000 | 1.5659 | | 0.0424 | 28500 | 1.5618 | | 0.0432 | 29000 | 1.5634 | | 0.0439 | 29500 | 1.562 | | 0.0447 | 30000 | 1.5646 | | 0.0454 | 30500 | 1.5815 | | 0.0461 | 31000 | 1.573 | | 0.0469 | 31500 | 1.5864 | | 0.0476 | 32000 | 1.5908 | | 0.0484 | 32500 | 1.5815 | | 0.0491 | 33000 | 1.5854 | | 0.0499 | 33500 | 1.5813 | | 0.0506 | 34000 | 1.6013 | | 0.0514 | 34500 | 1.594 | | 0.0521 | 35000 | 1.5883 | | 0.0528 | 35500 | 1.5985 | | 0.0536 | 36000 | 1.59 | | 0.0543 | 36500 | 1.5962 | | 0.0551 | 37000 | 1.596 | | 0.0558 | 37500 | 1.6023 | | 0.0566 | 38000 | 1.5883 | | 0.0573 | 38500 | 1.5868 | | 0.0581 | 39000 | 1.588 | | 0.0588 | 39500 | 1.5958 | | 0.0595 | 40000 | 1.5946 | | 0.0603 | 40500 | 1.586 | | 0.0610 | 41000 | 1.5867 | | 0.0618 | 41500 | 1.5967 | | 0.0625 | 42000 | 1.5965 | | 0.0633 | 42500 | 1.5946 | | 0.0640 | 43000 | 1.591 | | 0.0648 | 43500 | 1.5862 | | 0.0655 | 44000 | 1.5961 | | 0.0662 | 44500 | 1.5967 | | 0.0670 | 45000 | 1.5887 | | 0.0677 | 45500 | 1.5971 | | 0.0685 | 46000 | 1.5909 | | 0.0692 | 46500 | 1.5943 | | 0.0700 | 47000 | 1.5931 | | 0.0707 | 47500 | 1.6025 | | 0.0715 | 48000 | 1.5896 | | 0.0722 | 48500 | 1.5989 | | 0.0729 | 49000 | 1.6005 | | 0.0737 | 49500 | 1.5838 | | 0.0744 | 50000 | 1.592 | | 0.0752 | 50500 | 1.5937 | | 0.0759 | 51000 | 1.5943 | | 0.0767 | 51500 | 1.5971 | | 0.0774 | 52000 | 1.6036 | | 0.0782 | 52500 | 1.5846 | | 0.0789 | 53000 | 1.589 | | 0.0796 | 53500 | 1.6005 | | 0.0804 | 54000 | 1.5976 | | 0.0811 | 54500 | 1.6018 | | 0.0819 | 55000 | 1.6008 | | 0.0826 | 55500 | 1.6026 | | 0.0834 | 56000 | 1.5936 | | 0.0841 | 56500 | 1.5913 | | 0.0849 | 57000 | 1.5895 | | 0.0856 | 57500 | 1.5982 | | 0.0863 | 58000 | 1.5934 | | 0.0871 | 58500 | 1.595 | | 0.0878 | 59000 | 1.5959 | | 0.0886 | 59500 | 1.5908 | | 0.0893 | 60000 | 1.5944 | | 0.0901 | 60500 | 1.5875 | | 0.0908 | 61000 | 1.5999 | | 0.0916 | 61500 | 1.595 | | 0.0923 | 62000 | 1.5928 | | 0.0930 | 62500 | 1.5947 | | 0.0938 | 63000 | 1.5934 | | 0.0945 | 63500 | 1.5983 | | 0.0953 | 64000 | 1.5853 | | 0.0960 | 64500 | 1.5948 | | 0.0968 | 65000 | 1.6028 | | 0.0975 | 65500 | 1.6037 | | 0.0982 | 66000 | 1.59 | | 0.0990 | 66500 | 1.599 | | 0.0997 | 67000 | 1.5843 | | 0.0007 | 500 | 1.4423 | | 0.0015 | 1000 | 1.3901 | | 0.0022 | 1500 | 1.372 | | 0.0030 | 2000 | 1.3518 | | 0.0037 | 2500 | 1.3615 | | 0.0045 | 3000 | 1.3623 | | 0.0052 | 3500 | 1.3603 | | 0.0060 | 4000 | 1.3798 | | 0.0067 | 4500 | 1.3699 | | 0.0074 | 5000 | 1.3708 | | 0.0082 | 5500 | 1.378 | | 0.0089 | 6000 | 1.3783 | | 0.0097 | 6500 | 1.3851 | | 0.0104 | 7000 | 1.3839 | | 0.0112 | 7500 | 1.3898 | | 0.0119 | 8000 | 1.3933 | | 0.0127 | 8500 | 1.3982 | | 0.0134 | 9000 | 1.4051 | | 0.0141 | 9500 | 1.3944 | | 0.0149 | 10000 | 1.4093 | | 0.0156 | 10500 | 1.4077 | | 0.0164 | 11000 | 1.4264 | | 0.0171 | 11500 | 1.4322 | | 0.0179 | 12000 | 1.4286 | | 0.0186 | 12500 | 1.4325 | | 0.0194 | 13000 | 1.4267 | | 0.0201 | 13500 | 1.4299 | | 0.0208 | 14000 | 1.4403 | | 0.0216 | 14500 | 1.4497 | | 0.0223 | 15000 | 1.4386 | | 0.0231 | 15500 | 1.4528 | | 0.0238 | 16000 | 1.457 | | 0.0246 | 16500 | 1.4547 | | 0.0253 | 17000 | 1.4616 | | 0.0261 | 17500 | 1.4639 | | 0.0268 | 18000 | 1.4819 | | 0.0275 | 18500 | 1.4764 | | 0.0283 | 19000 | 1.469 | | 0.0290 | 19500 | 1.4792 | | 0.0298 | 20000 | 1.4916 | | 0.0305 | 20500 | 1.4885 | | 0.0313 | 21000 | 1.4985 | | 0.0320 | 21500 | 1.5038 | | 0.0327 | 22000 | 1.5118 | | 0.0335 | 22500 | 1.5023 | | 0.0342 | 23000 | 1.5129 | | 0.0350 | 23500 | 1.5197 | | 0.0357 | 24000 | 1.5174 | | 0.0365 | 24500 | 1.5276 | | 0.0372 | 25000 | 1.5172 | | 0.0380 | 25500 | 1.535 | | 0.0387 | 26000 | 1.545 | | 0.0394 | 26500 | 1.5388 | | 0.0402 | 27000 | 1.5461 | | 0.0409 | 27500 | 1.5492 | | 0.0417 | 28000 | 1.5533 | | 0.0424 | 28500 | 1.5527 | | 0.0432 | 29000 | 1.5537 | | 0.0439 | 29500 | 1.5545 | | 0.0447 | 30000 | 1.557 | | 0.0454 | 30500 | 1.5752 | | 0.0461 | 31000 | 1.5677 | | 0.0469 | 31500 | 1.5822 | | 0.0476 | 32000 | 1.5889 | | 0.0484 | 32500 | 1.5801 | | 0.0491 | 33000 | 1.5853 | | 0.0499 | 33500 | 1.5818 | | 0.0506 | 34000 | 1.6027 | | 0.0514 | 34500 | 1.5961 | | 0.0521 | 35000 | 1.5893 | | 0.0528 | 35500 | 1.5997 | | 0.0536 | 36000 | 1.5918 | | 0.0543 | 36500 | 1.598 | | 0.0551 | 37000 | 1.5974 | | 0.0558 | 37500 | 1.6028 | | 0.0566 | 38000 | 1.5893 | | 0.0573 | 38500 | 1.5878 | | 0.0581 | 39000 | 1.5889 | | 0.0588 | 39500 | 1.5971 | | 0.0595 | 40000 | 1.5965 | | 0.0603 | 40500 | 1.5868 | | 0.0610 | 41000 | 1.5868 | | 0.0618 | 41500 | 1.5979 | | 0.0625 | 42000 | 1.5976 | | 0.0633 | 42500 | 1.597 | | 0.0640 | 43000 | 1.593 | | 0.0648 | 43500 | 1.5869 | | 0.0655 | 44000 | 1.5971 | | 0.0662 | 44500 | 1.599 | | 0.0670 | 45000 | 1.5905 | | 0.0677 | 45500 | 1.5977 | | 0.0685 | 46000 | 1.5917 | | 0.0692 | 46500 | 1.5959 | | 0.0700 | 47000 | 1.5943 | | 0.0707 | 47500 | 1.6041 | | 0.0715 | 48000 | 1.5901 | | 0.0722 | 48500 | 1.6012 | | 0.0729 | 49000 | 1.6015 | | 0.0737 | 49500 | 1.5852 | | 0.0744 | 50000 | 1.5937 | | 0.0752 | 50500 | 1.5948 | | 0.0759 | 51000 | 1.5959 | | 0.0767 | 51500 | 1.5986 | | 0.0774 | 52000 | 1.6054 | | 0.0782 | 52500 | 1.5856 | | 0.0789 | 53000 | 1.59 | | 0.0796 | 53500 | 1.602 | | 0.0804 | 54000 | 1.6001 | | 0.0811 | 54500 | 1.6038 | | 0.0819 | 55000 | 1.6025 | | 0.0826 | 55500 | 1.6044 | | 0.0834 | 56000 | 1.5953 | | 0.0841 | 56500 | 1.5931 | | 0.0849 | 57000 | 1.5909 | | 0.0856 | 57500 | 1.5993 | | 0.0863 | 58000 | 1.5947 | | 0.0871 | 58500 | 1.5971 | | 0.0878 | 59000 | 1.5979 | | 0.0886 | 59500 | 1.5922 | | 0.0893 | 60000 | 1.5965 | | 0.0901 | 60500 | 1.5888 | | 0.0908 | 61000 | 1.6022 | | 0.0916 | 61500 | 1.596 | | 0.0923 | 62000 | 1.5939 | | 0.0930 | 62500 | 1.5958 | | 0.0938 | 63000 | 1.5948 | | 0.0945 | 63500 | 1.5997 | | 0.0953 | 64000 | 1.5853 | | 0.0960 | 64500 | 1.5963 | | 0.0968 | 65000 | 1.6045 | | 0.0975 | 65500 | 1.6053 | | 0.0982 | 66000 | 1.5907 | | 0.0990 | 66500 | 1.6007 | | 0.0997 | 67000 | 1.5852 | | 0.0007 | 500 | 1.3955 | | 0.0015 | 1000 | 1.3248 | | 0.0022 | 1500 | 1.2994 | | 0.0030 | 2000 | 1.2789 | | 0.0037 | 2500 | 1.2912 | | 0.0045 | 3000 | 1.2885 | | 0.0052 | 3500 | 1.289 | | 0.0060 | 4000 | 1.309 | | 0.0067 | 4500 | 1.3013 | | 0.0074 | 5000 | 1.3026 | | 0.0082 | 5500 | 1.3099 | | 0.0089 | 6000 | 1.3111 | | 0.0097 | 6500 | 1.3189 | | 0.0104 | 7000 | 1.3209 | | 0.0112 | 7500 | 1.325 | | 0.0119 | 8000 | 1.3299 | | 0.0127 | 8500 | 1.3355 | | 0.0134 | 9000 | 1.3452 | | 0.0141 | 9500 | 1.3335 | | 0.0149 | 10000 | 1.3486 | | 0.0156 | 10500 | 1.3515 | | 0.0164 | 11000 | 1.3718 | | 0.0171 | 11500 | 1.378 | | 0.0179 | 12000 | 1.3769 | | 0.0186 | 12500 | 1.3805 | | 0.0194 | 13000 | 1.375 | | 0.0201 | 13500 | 1.3805 | | 0.0208 | 14000 | 1.3915 | | 0.0216 | 14500 | 1.4036 | | 0.0223 | 15000 | 1.3923 | | 0.0231 | 15500 | 1.4076 | | 0.0238 | 16000 | 1.4129 | | 0.0246 | 16500 | 1.4123 | | 0.0253 | 17000 | 1.4217 | | 0.0261 | 17500 | 1.4263 | | 0.0268 | 18000 | 1.4456 | | 0.0275 | 18500 | 1.4411 | | 0.0283 | 19000 | 1.4338 | | 0.0290 | 19500 | 1.4467 | | 0.0298 | 20000 | 1.4591 | | 0.0305 | 20500 | 1.4578 | | 0.0313 | 21000 | 1.4697 | | 0.0320 | 21500 | 1.4747 | | 0.0327 | 22000 | 1.4856 | | 0.0335 | 22500 | 1.4763 | | 0.0342 | 23000 | 1.4889 | | 0.0350 | 23500 | 1.4964 | | 0.0357 | 24000 | 1.4953 | | 0.0365 | 24500 | 1.5073 | | 0.0372 | 25000 | 1.4992 | | 0.0380 | 25500 | 1.5179 | | 0.0387 | 26000 | 1.5303 | | 0.0394 | 26500 | 1.5236 | | 0.0402 | 27000 | 1.5326 | | 0.0409 | 27500 | 1.5361 | | 0.0417 | 28000 | 1.5408 | | 0.0424 | 28500 | 1.5432 | | 0.0432 | 29000 | 1.5443 | | 0.0439 | 29500 | 1.5472 | | 0.0447 | 30000 | 1.55 | | 0.0454 | 30500 | 1.5695 | | 0.0461 | 31000 | 1.5636 | | 0.0469 | 31500 | 1.5786 | | 0.0476 | 32000 | 1.5875 | | 0.0484 | 32500 | 1.5791 | | 0.0491 | 33000 | 1.5855 | | 0.0499 | 33500 | 1.5828 | | 0.0506 | 34000 | 1.6046 | | 0.0514 | 34500 | 1.5983 | | 0.0521 | 35000 | 1.5909 | | 0.0528 | 35500 | 1.6011 | | 0.0536 | 36000 | 1.5943 | | 0.0543 | 36500 | 1.6 | | 0.0551 | 37000 | 1.5998 | | 0.0558 | 37500 | 1.6046 | | 0.0566 | 38000 | 1.5907 | | 0.0573 | 38500 | 1.5892 | | 0.0581 | 39000 | 1.5909 | | 0.0588 | 39500 | 1.5989 | | 0.0595 | 40000 | 1.599 | | 0.0603 | 40500 | 1.5886 | | 0.0610 | 41000 | 1.5879 | | 0.0618 | 41500 | 1.5998 | | 0.0625 | 42000 | 1.5994 | | 0.0633 | 42500 | 1.5993 | | 0.0640 | 43000 | 1.596 | | 0.0648 | 43500 | 1.5884 | | 0.0655 | 44000 | 1.5987 | | 0.0662 | 44500 | 1.6018 | | 0.0670 | 45000 | 1.5936 | | 0.0677 | 45500 | 1.5991 | | 0.0685 | 46000 | 1.5933 | | 0.0692 | 46500 | 1.5984 | | 0.0700 | 47000 | 1.5964 | | 0.0707 | 47500 | 1.6062 | | 0.0715 | 48000 | 1.592 | | 0.0722 | 48500 | 1.6034 | | 0.0729 | 49000 | 1.6034 | | 0.0737 | 49500 | 1.5871 | | 0.0744 | 50000 | 1.5959 | | 0.0752 | 50500 | 1.5971 | | 0.0759 | 51000 | 1.5983 | | 0.0767 | 51500 | 1.6004 | | 0.0774 | 52000 | 1.6079 | | 0.0782 | 52500 | 1.5875 | | 0.0789 | 53000 | 1.5921 | | 0.0796 | 53500 | 1.6043 | | 0.0804 | 54000 | 1.6029 | | 0.0811 | 54500 | 1.606 | | 0.0819 | 55000 | 1.6049 | | 0.0826 | 55500 | 1.6068 | | 0.0834 | 56000 | 1.5973 | | 0.0841 | 56500 | 1.5953 | | 0.0849 | 57000 | 1.593 | | 0.0856 | 57500 | 1.6007 | | 0.0863 | 58000 | 1.5968 | | 0.0871 | 58500 | 1.5996 | | 0.0878 | 59000 | 1.6006 | | 0.0886 | 59500 | 1.5944 | | 0.0893 | 60000 | 1.5995 | | 0.0901 | 60500 | 1.5913 | | 0.0908 | 61000 | 1.6049 | | 0.0916 | 61500 | 1.5977 | | 0.0923 | 62000 | 1.5956 | | 0.0930 | 62500 | 1.5974 | | 0.0938 | 63000 | 1.5975 | | 0.0945 | 63500 | 1.6016 | | 0.0953 | 64000 | 1.5866 | | 0.0960 | 64500 | 1.5987 | | 0.0968 | 65000 | 1.6069 | | 0.0975 | 65500 | 1.6074 | | 0.0982 | 66000 | 1.592 | | 0.0990 | 66500 | 1.6026 | | 0.0997 | 67000 | 1.5871 | </details> ### Framework Versions - Python: 3.11.11 - Sentence Transformers: 3.4.1 - Transformers: 4.49.0 - PyTorch: 2.6.0+cu124 - Accelerate: 1.5.2 - Datasets: 3.4.1 - Tokenizers: 0.21.1 ## Citation ### BibTeX #### Sentence Transformers and SoftmaxLoss ```bibtex @inproceedings{reimers-2019-sentence-bert, title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks", author = "Reimers, Nils and Gurevych, Iryna", booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing", month = "11", year = "2019", publisher = "Association for Computational Linguistics", url = "https://arxiv.org/abs/1908.10084", } ``` #### MultipleNegativesRankingLoss ```bibtex @misc{henderson2017efficient, title={Efficient Natural Language Response Suggestion for Smart Reply}, author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil}, year={2017}, eprint={1705.00652}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` <!-- ## Glossary *Clearly define terms in order to be accessible across audiences.* --> <!-- ## Model Card Authors *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.* --> <!-- ## Model Card Contact *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.* -->
iamjoshgreen/iamkora
iamjoshgreen
2025-03-25T18:17:15Z
0
0
diffusers
[ "diffusers", "flux", "lora", "replicate", "text-to-image", "en", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "license:other", "region:us" ]
text-to-image
2025-03-25T18:07:11Z
--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md language: - en tags: - flux - diffusers - lora - replicate base_model: "black-forest-labs/FLUX.1-dev" pipeline_tag: text-to-image # widget: # - text: >- # prompt # output: # url: https://... instance_prompt: iamkora --- # Iamkora <Gallery /> Trained on Replicate using: https://replicate.com/ostris/flux-dev-lora-trainer/train ## Trigger words You should use `iamkora` to trigger the image generation. ## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers) ```py from diffusers import AutoPipelineForText2Image import torch pipeline = AutoPipelineForText2Image.from_pretrained('black-forest-labs/FLUX.1-dev', torch_dtype=torch.float16).to('cuda') pipeline.load_lora_weights('iamjoshgreen/iamkora', weight_name='lora.safetensors') image = pipeline('your prompt').images[0] ``` For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters)
natsudalkr/test-v3
natsudalkr
2025-03-25T18:17:13Z
0
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "trl", "sft", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "4-bit", "bitsandbytes", "region:us" ]
text-generation
2025-03-25T15:48:21Z
--- library_name: transformers tags: - trl - sft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
LHRuig/darrsx
LHRuig
2025-03-25T18:16:42Z
0
0
diffusers
[ "diffusers", "safetensors", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T18:16:23Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: darrsx --- # darrsx <Gallery /> ## Model description darrsx lora ## Trigger words You should use `darrsx` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/darrsx/tree/main) them in the Files & versions tab.
lesso11/47b05e4e-9c6c-40e4-93c5-812408aee76a
lesso11
2025-03-25T18:15:45Z
0
0
peft
[ "peft", "safetensors", "llama", "axolotl", "generated_from_trainer", "custom_code", "base_model:NousResearch/Yarn-Solar-10b-32k", "base_model:adapter:NousResearch/Yarn-Solar-10b-32k", "license:apache-2.0", "region:us" ]
null
2025-03-25T13:22:07Z
--- library_name: peft license: apache-2.0 base_model: NousResearch/Yarn-Solar-10b-32k tags: - axolotl - generated_from_trainer model-index: - name: 47b05e4e-9c6c-40e4-93c5-812408aee76a results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.4.1` ```yaml adapter: lora base_model: NousResearch/Yarn-Solar-10b-32k bf16: auto chat_template: llama3 dataset_prepared_path: null datasets: - data_files: - 65d9e80afe69aff1_train_data.json ds_type: json format: custom path: /workspace/input_data/65d9e80afe69aff1_train_data.json type: field_input: documents field_instruction: question field_output: answer format: '{instruction} {input}' no_input_format: '{instruction}' system_format: '{system}' system_prompt: '' debug: null deepspeed: null do_eval: true early_stopping_patience: 3 eval_batch_size: 4 eval_max_new_tokens: 128 eval_steps: 500 evals_per_epoch: null flash_attention: true fp16: false fsdp: null fsdp_config: null gradient_accumulation_steps: 8 gradient_checkpointing: true group_by_length: true hub_model_id: lesso11/47b05e4e-9c6c-40e4-93c5-812408aee76a hub_repo: null hub_strategy: checkpoint hub_token: null learning_rate: 0.000211 load_in_4bit: false load_in_8bit: false local_rank: null logging_steps: 50 lora_alpha: 128 lora_dropout: 0.15 lora_fan_in_fan_out: null lora_model_dir: null lora_r: 64 lora_target_linear: true lr_scheduler: cosine max_grad_norm: 1.0 max_steps: 500 micro_batch_size: 4 mlflow_experiment_name: /tmp/65d9e80afe69aff1_train_data.json model_type: AutoModelForCausalLM num_epochs: 10 optimizer: adamw_torch_fused output_dir: miner_id_24 pad_to_sequence_len: true resume_from_checkpoint: null s2_attention: null sample_packing: false save_steps: 500 saves_per_epoch: null seed: 110 sequence_len: 1024 special_tokens: pad_token: </s> strict: false tf32: true tokenizer_type: AutoTokenizer train_on_inputs: false trust_remote_code: true val_set_size: 0.05 wandb_entity: null wandb_mode: online wandb_name: a1a050a4-6a01-49dd-9cd7-289119b180f3 wandb_project: 11a wandb_run: your_name wandb_runid: a1a050a4-6a01-49dd-9cd7-289119b180f3 warmup_steps: 100 weight_decay: 0.0 xformers_attention: null ``` </details><br> # 47b05e4e-9c6c-40e4-93c5-812408aee76a This model is a fine-tuned version of [NousResearch/Yarn-Solar-10b-32k](https://huggingface.co/NousResearch/Yarn-Solar-10b-32k) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.0285 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.000211 - train_batch_size: 4 - eval_batch_size: 4 - seed: 110 - gradient_accumulation_steps: 8 - total_train_batch_size: 32 - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 100 - training_steps: 500 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:------:|:----:|:---------------:| | No log | 0.0008 | 1 | 1.6909 | | 8.2657 | 0.4238 | 500 | 1.0285 | ### Framework versions - PEFT 0.13.2 - Transformers 4.46.0 - Pytorch 2.5.0+cu124 - Datasets 3.0.1 - Tokenizers 0.20.1
duchao1210/16_bit_kmap_gguf_3epoch_2k_kmap_queen3B_solution
duchao1210
2025-03-25T18:15:10Z
0
0
transformers
[ "transformers", "gguf", "qwen2", "text-generation-inference", "unsloth", "en", "license:apache-2.0", "endpoints_compatible", "region:us", "conversational" ]
null
2025-03-25T18:13:31Z
--- base_model: unsloth/qwen2.5-3b-instruct-unsloth-bnb-4bit tags: - text-generation-inference - transformers - unsloth - qwen2 - gguf license: apache-2.0 language: - en --- # Uploaded model - **Developed by:** duchao1210 - **License:** apache-2.0 - **Finetuned from model :** unsloth/qwen2.5-3b-instruct-unsloth-bnb-4bit This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
LHRuig/brandzarsx
LHRuig
2025-03-25T18:13:44Z
0
0
diffusers
[ "diffusers", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T18:13:26Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: brandzarsx --- # brandzarsx <Gallery /> ## Model description brandzarsx lora ## Trigger words You should use `brandzarsx` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/brandzarsx/tree/main) them in the Files & versions tab.
Bagratuni/dare2_hnf
Bagratuni
2025-03-25T18:13:33Z
0
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "feature-extraction", "arxiv:1910.09700", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2025-03-25T17:59:59Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
YhangChen/DeepSeek-R1-Distill-Qwen-1.5B-GRPO
YhangChen
2025-03-25T18:13:21Z
0
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "generated_from_trainer", "open-r1", "trl", "grpo", "conversational", "dataset:YhangChen/math_train_5000", "arxiv:2402.03300", "base_model:deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B", "base_model:finetune:deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-24T10:44:48Z
--- base_model: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B datasets: YhangChen/math_train_5000 library_name: transformers model_name: DeepSeek-R1-Distill-Qwen-1.5B-GRPO tags: - generated_from_trainer - open-r1 - trl - grpo licence: license --- # Model Card for DeepSeek-R1-Distill-Qwen-1.5B-GRPO This model is a fine-tuned version of [deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B) on the [YhangChen/math_train_5000](https://huggingface.co/datasets/YhangChen/math_train_5000) dataset. It has been trained using [TRL](https://github.com/huggingface/trl). ## Quick start ```python from transformers import pipeline question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?" generator = pipeline("text-generation", model="YhangChen/DeepSeek-R1-Distill-Qwen-1.5B-GRPO", device="cuda") output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0] print(output["generated_text"]) ``` ## Training procedure [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/yhangchen/openr1/runs/vtu6ebkd) This model was trained with GRPO, a method introduced in [DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models](https://huggingface.co/papers/2402.03300). ### Framework versions - TRL: 0.16.0.dev0 - Transformers: 4.49.0 - Pytorch: 2.5.1 - Datasets: 3.4.1 - Tokenizers: 0.21.1 ## Citations Cite GRPO as: ```bibtex @article{zhihong2024deepseekmath, title = {{DeepSeekMath: Pushing the Limits of Mathematical Reasoning in Open Language Models}}, author = {Zhihong Shao and Peiyi Wang and Qihao Zhu and Runxin Xu and Junxiao Song and Mingchuan Zhang and Y. K. Li and Y. Wu and Daya Guo}, year = 2024, eprint = {arXiv:2402.03300}, } ``` Cite TRL as: ```bibtex @misc{vonwerra2022trl, title = {{TRL: Transformer Reinforcement Learning}}, author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec}, year = 2020, journal = {GitHub repository}, publisher = {GitHub}, howpublished = {\url{https://github.com/huggingface/trl}} } ```
LHRuig/elnazshakersx
LHRuig
2025-03-25T18:13:06Z
0
0
diffusers
[ "diffusers", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T18:12:48Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: elnazshakersx --- # elnazshakersx <Gallery /> ## Model description elnazshakersx lora ## Trigger words You should use `elnazshakersx` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/elnazshakersx/tree/main) them in the Files & versions tab.
LHRuig/joncareersx
LHRuig
2025-03-25T18:11:08Z
0
0
diffusers
[ "diffusers", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T18:10:48Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: joncareersx --- # joncareersx <Gallery /> ## Model description joncareersx lora ## Trigger words You should use `joncareersx` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/joncareersx/tree/main) them in the Files & versions tab.
LHRuig/benpinksx
LHRuig
2025-03-25T18:10:35Z
0
0
diffusers
[ "diffusers", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T18:10:17Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: benpinksx --- # benpinksx <Gallery /> ## Model description benpinksx lora ## Trigger words You should use `benpinksx` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/benpinksx/tree/main) them in the Files & versions tab.
kamel-usp/jbcs2025_bertimbau_base-C5
kamel-usp
2025-03-25T18:09:51Z
2
0
null
[ "safetensors", "bert", "aes", "pt", "en", "dataset:kamel-usp/aes_enem_dataset", "base_model:neuralmind/bert-base-portuguese-cased", "base_model:finetune:neuralmind/bert-base-portuguese-cased", "model-index", "region:us" ]
null
2025-03-16T01:27:24Z
--- language: - pt - en tags: - aes datasets: - kamel-usp/aes_enem_dataset base_model: neuralmind/bert-base-portuguese-cased metrics: - accuracy - qwk model-index: - name: bertimbau_base-C5 results: - task: type: text-classification name: Automated Essay Score dataset: name: Automated Essay Score ENEM Dataset type: kamel-usp/aes_enem_dataset config: JBCS2025 split: test metrics: - name: Macro F1 type: f1 value: 0.2055897809038726 - name: QWK type: qwk value: 0.476219483623073 - name: Weighted Macro F1 type: f1 value: 0.25808413038205613 --- # Model ID: bertimbau_base-C5 ## Results | | test_data | |:-----------------|------------:| | eval_accuracy | 0.318841 | | eval_RMSE | 61.2905 | | eval_QWK | 0.476219 | | eval_Macro_F1 | 0.20559 | | eval_Weighted_F1 | 0.258084 | | eval_Micro_F1 | 0.318841 | | eval_HDIV | 0.130435 |
LHRuig/elonmuskksx
LHRuig
2025-03-25T18:09:32Z
0
0
diffusers
[ "diffusers", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T18:09:14Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: elonmuskksx --- # elonmuskksx <Gallery /> ## Model description elonmuskksx lora ## Trigger words You should use `elonmuskksx` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/elonmuskksx/tree/main) them in the Files & versions tab.
LHRuig/nehasx
LHRuig
2025-03-25T18:08:54Z
0
0
diffusers
[ "diffusers", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T18:08:36Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: nehasx --- # nehasx <Gallery /> ## Model description nehasx lora ## Trigger words You should use `nehasx` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/nehasx/tree/main) them in the Files & versions tab.
kamel-usp/jbcs2025_bertimbau_base-C4
kamel-usp
2025-03-25T18:08:19Z
2
0
null
[ "safetensors", "bert", "aes", "pt", "en", "dataset:kamel-usp/aes_enem_dataset", "base_model:neuralmind/bert-base-portuguese-cased", "base_model:finetune:neuralmind/bert-base-portuguese-cased", "model-index", "region:us" ]
null
2025-03-16T01:24:19Z
--- language: - pt - en tags: - aes datasets: - kamel-usp/aes_enem_dataset base_model: neuralmind/bert-base-portuguese-cased metrics: - accuracy - qwk model-index: - name: bertimbau_base-C4 results: - task: type: text-classification name: Automated Essay Score dataset: name: Automated Essay Score ENEM Dataset type: kamel-usp/aes_enem_dataset config: JBCS2025 split: test metrics: - name: Macro F1 type: f1 value: 0.36114488348530904 - name: QWK type: qwk value: 0.6258134490238612 - name: Weighted Macro F1 type: f1 value: 0.6545879036165807 --- # Model ID: bertimbau_base-C4 ## Results | | test_data | |:-----------------|------------:| | eval_accuracy | 0.644928 | | eval_RMSE | 26.3752 | | eval_QWK | 0.625813 | | eval_Macro_F1 | 0.361145 | | eval_Weighted_F1 | 0.654588 | | eval_Micro_F1 | 0.644928 | | eval_HDIV | 0.00724638 |
kamel-usp/jbcs2025_bertimbau_base-C2
kamel-usp
2025-03-25T18:05:48Z
2
0
null
[ "safetensors", "bert", "aes", "pt", "en", "dataset:kamel-usp/aes_enem_dataset", "base_model:neuralmind/bert-base-portuguese-cased", "base_model:finetune:neuralmind/bert-base-portuguese-cased", "model-index", "region:us" ]
null
2025-03-16T01:18:20Z
--- language: - pt - en tags: - aes datasets: - kamel-usp/aes_enem_dataset base_model: neuralmind/bert-base-portuguese-cased metrics: - accuracy - qwk model-index: - name: bertimbau_base-C2 results: - task: type: text-classification name: Automated Essay Score dataset: name: Automated Essay Score ENEM Dataset type: kamel-usp/aes_enem_dataset config: JBCS2025 split: test metrics: - name: Macro F1 type: f1 value: 0.27254317053298555 - name: QWK type: qwk value: 0.41025641025641035 - name: Weighted Macro F1 type: f1 value: 0.37216098145030935 --- # Model ID: bertimbau_base-C2 ## Results | | test_data | |:-----------------|------------:| | eval_accuracy | 0.369565 | | eval_RMSE | 55.7427 | | eval_QWK | 0.410256 | | eval_Macro_F1 | 0.272543 | | eval_Weighted_F1 | 0.372161 | | eval_Micro_F1 | 0.369565 | | eval_HDIV | 0.0652174 |
RayneAmes/kokujin8
RayneAmes
2025-03-25T18:02:53Z
0
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-25T16:58:17Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
kamel-usp/jbcs2025_bertimbau-large-C5
kamel-usp
2025-03-25T18:02:03Z
4
0
null
[ "safetensors", "bert", "aes", "pt", "en", "dataset:kamel-usp/aes_enem_dataset", "base_model:neuralmind/bert-large-portuguese-cased", "base_model:finetune:neuralmind/bert-large-portuguese-cased", "model-index", "region:us" ]
null
2025-03-16T01:05:44Z
--- language: - pt - en tags: - aes datasets: - kamel-usp/aes_enem_dataset base_model: neuralmind/bert-large-portuguese-cased metrics: - accuracy - qwk model-index: - name: bertimbau-large-C5 results: - task: type: text-classification name: Automated Essay Score dataset: name: Automated Essay Score ENEM Dataset type: kamel-usp/aes_enem_dataset config: JBCS2025 split: test metrics: - name: Macro F1 type: f1 value: 0.32328719577491555 - name: QWK type: qwk value: 0.4830137751303053 - name: Weighted Macro F1 type: f1 value: 0.3502415018406545 --- # Model ID: bertimbau-large-C5 ## Results | | test_data | |:-----------------|------------:| | eval_accuracy | 0.362319 | | eval_RMSE | 61.101 | | eval_QWK | 0.483014 | | eval_Macro_F1 | 0.323287 | | eval_Weighted_F1 | 0.350242 | | eval_Micro_F1 | 0.362319 | | eval_HDIV | 0.144928 |
souging/8e6b92c1-26e7-406f-80b8-fc857e196040
souging
2025-03-25T18:01:12Z
0
0
peft
[ "peft", "safetensors", "llama", "axolotl", "generated_from_trainer", "base_model:MLP-KTLim/llama-3-Korean-Bllossom-8B", "base_model:adapter:MLP-KTLim/llama-3-Korean-Bllossom-8B", "license:llama3", "region:us" ]
null
2025-03-25T17:35:48Z
--- library_name: peft license: llama3 base_model: MLP-KTLim/llama-3-Korean-Bllossom-8B tags: - axolotl - generated_from_trainer model-index: - name: 8e6b92c1-26e7-406f-80b8-fc857e196040 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.4.1` ```yaml adapter: lora base_model: MLP-KTLim/llama-3-Korean-Bllossom-8B bf16: auto dataset_prepared_path: null datasets: - data_files: - fdd02748a5c77af5_train_data.json ds_type: json format: custom path: /root/G.O.D-test/core/data/fdd02748a5c77af5_train_data.json type: field_instruction: func_before field_output: func_after format: '{instruction}' no_input_format: '{instruction}' system_format: '{system}' system_prompt: '' debug: null deepspeed: null eval_max_new_tokens: 128 eval_steps: 0 evals_per_epoch: null flash_attention: true fp16: null fsdp: null fsdp_config: null gradient_accumulation_steps: 4 gradient_checkpointing: false group_by_length: false hub_model_id: souging/8e6b92c1-26e7-406f-80b8-fc857e196040 hub_repo: null hub_strategy: checkpoint hub_token: null learning_rate: 0.000202 load_in_4bit: false load_in_8bit: false local_rank: null logging_steps: 1 lora_alpha: 64 lora_dropout: 0.05 lora_fan_in_fan_out: null lora_model_dir: null lora_r: 32 lora_target_linear: true lr_scheduler: cosine max_steps: 500 micro_batch_size: 2 mlflow_experiment_name: /tmp/fdd02748a5c77af5_train_data.json model_type: AutoModelForCausalLM num_epochs: 10 optimizer: adamw_bnb_8bit output_dir: miner_id_24 pad_to_sequence_len: false resume_from_checkpoint: null s2_attention: null sample_packing: false save_steps: 0 saves_per_epoch: null sequence_len: 1408 strict: false tf32: false tokenizer_type: AutoTokenizer train_on_inputs: false trust_remote_code: true wandb_entity: null wandb_mode: online wandb_name: d7b13495-a771-4013-96ef-d42f1ab3fedc wandb_project: Gradients-On-Demand wandb_run: your_name wandb_runid: d7b13495-a771-4013-96ef-d42f1ab3fedc warmup_steps: 100 weight_decay: 0.01 xformers_attention: null ``` </details><br> # 8e6b92c1-26e7-406f-80b8-fc857e196040 This model is a fine-tuned version of [MLP-KTLim/llama-3-Korean-Bllossom-8B](https://huggingface.co/MLP-KTLim/llama-3-Korean-Bllossom-8B) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.000202 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - distributed_type: multi-GPU - num_devices: 8 - gradient_accumulation_steps: 4 - total_train_batch_size: 64 - total_eval_batch_size: 16 - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 100 - training_steps: 500 ### Training results ### Framework versions - PEFT 0.13.2 - Transformers 4.46.0 - Pytorch 2.5.0+cu124 - Datasets 3.0.1 - Tokenizers 0.20.3
Bagratuni/dare4_hnf
Bagratuni
2025-03-25T17:59:57Z
0
0
transformers
[ "transformers", "safetensors", "xlm-roberta", "feature-extraction", "arxiv:1910.09700", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2025-03-25T17:48:10Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
genki10/BERT_AugV8_k5_task1_organization_sp040_lw030_fold4
genki10
2025-03-25T17:59:36Z
0
0
transformers
[ "transformers", "pytorch", "bert", "text-classification", "generated_from_trainer", "base_model:google-bert/bert-base-uncased", "base_model:finetune:google-bert/bert-base-uncased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2025-03-25T17:46:49Z
--- library_name: transformers license: apache-2.0 base_model: bert-base-uncased tags: - generated_from_trainer model-index: - name: BERT_AugV8_k5_task1_organization_sp040_lw030_fold4 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # BERT_AugV8_k5_task1_organization_sp040_lw030_fold4 This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6324 - Qwk: 0.5277 - Mse: 0.6324 - Rmse: 0.7952 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - num_epochs: 150 ### Training results | Training Loss | Epoch | Step | Validation Loss | Qwk | Mse | Rmse | |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:| | No log | 1.0 | 4 | 9.1512 | 0.0018 | 9.1512 | 3.0251 | | No log | 2.0 | 8 | 5.4836 | 0.0444 | 5.4836 | 2.3417 | | No log | 3.0 | 12 | 2.9584 | 0.0118 | 2.9584 | 1.7200 | | No log | 4.0 | 16 | 1.5845 | 0.0445 | 1.5845 | 1.2588 | | No log | 5.0 | 20 | 1.0553 | 0.0420 | 1.0553 | 1.0273 | | No log | 6.0 | 24 | 1.5129 | 0.0342 | 1.5129 | 1.2300 | | No log | 7.0 | 28 | 0.9238 | 0.0809 | 0.9238 | 0.9611 | | No log | 8.0 | 32 | 1.0810 | 0.1426 | 1.0810 | 1.0397 | | No log | 9.0 | 36 | 1.5196 | 0.2050 | 1.5196 | 1.2327 | | No log | 10.0 | 40 | 0.9072 | 0.2371 | 0.9072 | 0.9525 | | No log | 11.0 | 44 | 0.8024 | 0.4533 | 0.8024 | 0.8958 | | No log | 12.0 | 48 | 0.6978 | 0.4127 | 0.6978 | 0.8353 | | No log | 13.0 | 52 | 0.6533 | 0.4979 | 0.6533 | 0.8083 | | No log | 14.0 | 56 | 0.5802 | 0.5037 | 0.5802 | 0.7617 | | No log | 15.0 | 60 | 0.6725 | 0.5358 | 0.6725 | 0.8201 | | No log | 16.0 | 64 | 0.6115 | 0.5456 | 0.6115 | 0.7820 | | No log | 17.0 | 68 | 0.6413 | 0.5655 | 0.6413 | 0.8008 | | No log | 18.0 | 72 | 0.7044 | 0.5561 | 0.7044 | 0.8393 | | No log | 19.0 | 76 | 0.7050 | 0.5657 | 0.7050 | 0.8396 | | No log | 20.0 | 80 | 0.7654 | 0.5599 | 0.7654 | 0.8749 | | No log | 21.0 | 84 | 1.3086 | 0.3697 | 1.3086 | 1.1439 | | No log | 22.0 | 88 | 0.8007 | 0.5136 | 0.8007 | 0.8948 | | No log | 23.0 | 92 | 2.6553 | 0.1465 | 2.6553 | 1.6295 | | No log | 24.0 | 96 | 0.6583 | 0.5278 | 0.6583 | 0.8114 | | No log | 25.0 | 100 | 0.7613 | 0.5154 | 0.7613 | 0.8726 | | No log | 26.0 | 104 | 0.8989 | 0.4838 | 0.8989 | 0.9481 | | No log | 27.0 | 108 | 1.4490 | 0.3387 | 1.4490 | 1.2037 | | No log | 28.0 | 112 | 0.6986 | 0.5598 | 0.6986 | 0.8358 | | No log | 29.0 | 116 | 0.9816 | 0.4262 | 0.9816 | 0.9907 | | No log | 30.0 | 120 | 0.6020 | 0.5624 | 0.6020 | 0.7759 | | No log | 31.0 | 124 | 1.1014 | 0.3666 | 1.1014 | 1.0495 | | No log | 32.0 | 128 | 0.6093 | 0.5277 | 0.6093 | 0.7806 | | No log | 33.0 | 132 | 0.9285 | 0.4291 | 0.9285 | 0.9636 | | No log | 34.0 | 136 | 0.6324 | 0.5277 | 0.6324 | 0.7952 | ### Framework versions - Transformers 4.47.0 - Pytorch 2.5.1+cu121 - Datasets 3.3.1 - Tokenizers 0.21.0
RichardErkhov/sethuiyer_-_SynthIQ-7b-8bits
RichardErkhov
2025-03-25T17:59:10Z
0
0
null
[ "safetensors", "mistral", "8-bit", "bitsandbytes", "region:us" ]
null
2025-03-25T17:53:37Z
Quantization made by Richard Erkhov. [Github](https://github.com/RichardErkhov) [Discord](https://discord.gg/pvy7H8DZMG) [Request more models](https://github.com/RichardErkhov/quant_request) SynthIQ-7b - bnb 8bits - Model creator: https://huggingface.co/sethuiyer/ - Original model: https://huggingface.co/sethuiyer/SynthIQ-7b/ Original model description: --- language: - en license: llama2 library_name: transformers tags: - mistral - merge datasets: - stingning/ultrachat - garage-bAInd/Open-Platypus - Open-Orca/OpenOrca - TIGER-Lab/MathInstruct - OpenAssistant/oasst_top1_2023-08-25 - teknium/openhermes - meta-math/MetaMathQA - Open-Orca/SlimOrca pipeline_tag: text-generation base_model: - Weyaxi/OpenHermes-2.5-neural-chat-v3-3-openchat-3.5-1210-Slerp - ehartford/dolphin-2.1-mistral-7b - Open-Orca/Mistral-7B-OpenOrca - bhenrym14/mistral-7b-platypus-fp16 - ehartford/samantha-1.2-mistral-7b - iteknium/CollectiveCognition-v1.1-Mistral-7B - HuggingFaceH4/zephyr-7b-alpha model-index: - name: sethuiyer/SynthIQ-7b results: - task: type: text-generation name: Text Generation dataset: name: AI2 Reasoning Challenge (25-Shot) type: ai2_arc config: ARC-Challenge split: test args: num_few_shot: 25 metrics: - type: acc_norm value: 65.87 name: normalized accuracy source: url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=sethuiyer/SynthIQ-7b name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: HellaSwag (10-Shot) type: hellaswag split: validation args: num_few_shot: 10 metrics: - type: acc_norm value: 85.82 name: normalized accuracy source: url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=sethuiyer/SynthIQ-7b name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: MMLU (5-Shot) type: cais/mmlu config: all split: test args: num_few_shot: 5 metrics: - type: acc value: 64.75 name: accuracy source: url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=sethuiyer/SynthIQ-7b name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: TruthfulQA (0-shot) type: truthful_qa config: multiple_choice split: validation args: num_few_shot: 0 metrics: - type: mc2 value: 57 source: url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=sethuiyer/SynthIQ-7b name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: Winogrande (5-shot) type: winogrande config: winogrande_xl split: validation args: num_few_shot: 5 metrics: - type: acc value: 78.69 name: accuracy source: url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=sethuiyer/SynthIQ-7b name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: GSM8k (5-shot) type: gsm8k config: main split: test args: num_few_shot: 5 metrics: - type: acc value: 64.06 name: accuracy source: url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=sethuiyer/SynthIQ-7b name: Open LLM Leaderboard --- <p align="center"> <img src="https://codeberg.org/aninokuma/DeydooAssistant/raw/branch/main/logo.webp" height="256px" alt="SynthIQ"> </p> # SynthIQ This is SynthIQ, rated **92.23/100** by GPT-4 across varied complex prompts. I used [mergekit](https://github.com/cg123/mergekit) to merge models. | Benchmark Name | Score | | ---------------- | ----- | | ARC | 65.87 | | HellaSwag | 85.82 | | MMLU | 64.75 | | TruthfulQA | 57.00 | | Winogrande | 78.69 | | GSM8K | 64.06 | | AGIEval | 42.67 | | GPT4All | 73.71 | | Bigbench | 44.59 | ## Update - 19/01/2024 Tested to work well with autogen and CrewAI GGUF Files [Q4_K_M](https://huggingface.co/sethuiyer/SynthIQ_GGUF/blob/main/synthiq.Q4_K_M.gguf) - medium, balanced quality - recommended [Q_6_K](https://huggingface.co/sethuiyer/SynthIQ_GGUF/blob/main/synthiq.Q6_K.gguf) - very large, extremely low quality loss [Q8_0](https://huggingface.co/sethuiyer/SynthIQ_GGUF/blob/main/synthiq.Q8.gguf) - very large, extremely low quality loss - not recommended **Important Update**: SynthIQ is now available on Ollama. You can use it by running the command ```ollama run stuehieyr/synthiq``` in your terminal. If you have limited computing resources, check out this [video](https://www.youtube.com/watch?v=Qa1h7ygwQq8) to learn how to run it on a Google Colab backend. # Yaml Config ```yaml slices: - sources: - model: Weyaxi/OpenHermes-2.5-neural-chat-v3-3-openchat-3.5-1210-Slerp layer_range: [0, 32] - model: uukuguy/speechless-mistral-six-in-one-7b layer_range: [0, 32] merge_method: slerp base_model: mistralai/Mistral-7B-v0.1 parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 # fallback for rest of tensors tokenizer_source: union dtype: bfloat16 ``` <!-- prompt-template start --> ## Prompt template: ChatML ``` <|im_start|>system {system_message}<|im_end|> <|im_start|>user {prompt}<|im_end|> <|im_start|>assistant ``` <!-- prompt-template end --> License is LLama2 license as uukuguy/speechless-mistral-six-in-one-7b is llama2 license. # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard) Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_sethuiyer__SynthIQ-7b) # [Nous Benchmark Evalation Results](https://huggingface.co/spaces/mlabonne/Yet_Another_LLM_Leaderboard) Detailed results can be found [here](https://gist.github.com/sethuiyer/f47dee388a4e95d46181c98d37d66a58) # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard) Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_sethuiyer__SynthIQ-7b) | Metric |Value| |---------------------------------|----:| |Avg. |69.37| |AI2 Reasoning Challenge (25-Shot)|65.87| |HellaSwag (10-Shot) |85.82| |MMLU (5-Shot) |64.75| |TruthfulQA (0-shot) |57.00| |Winogrande (5-shot) |78.69| |GSM8k (5-shot) |64.06|
Dortp58/Llava-11B-finetuned
Dortp58
2025-03-25T17:59:05Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2025-03-25T17:58:57Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
LHRuig/bloodbornesx
LHRuig
2025-03-25T17:58:58Z
0
0
diffusers
[ "diffusers", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T17:58:40Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: bloodbornesx --- # bloodbornesx <Gallery /> ## Model description bloodbornesx lora ## Trigger words You should use `bloodbornesx` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/bloodbornesx/tree/main) them in the Files & versions tab.
LHRuig/yasinsx
LHRuig
2025-03-25T17:58:22Z
0
0
diffusers
[ "diffusers", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T17:58:03Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: yasinsx --- # yasinsx <Gallery /> ## Model description yasinsx lora ## Trigger words You should use `yasinsx` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/yasinsx/tree/main) them in the Files & versions tab.
Deepanshu7284/BMU_LEGAL_SUMMARIZER_HYBRID_ILC
Deepanshu7284
2025-03-25T17:57:20Z
0
0
transformers
[ "transformers", "safetensors", "led", "text2text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2025-03-25T17:56:19Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
decube/bge-m3-sql
decube
2025-03-25T17:57:13Z
0
0
null
[ "safetensors", "xlm-roberta", "base_model:BAAI/bge-m3", "base_model:finetune:BAAI/bge-m3", "region:us" ]
null
2025-03-25T05:58:45Z
--- base_model: - BAAI/bge-m3 --- # BGE M3 SQL Finetune [BAAI/bge-m3](https://huggingface.co/BAAI/bge-m3) on SQL context.
LHRuig/vladimirssx
LHRuig
2025-03-25T17:57:11Z
0
0
diffusers
[ "diffusers", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T17:56:53Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: vladimirsx --- # vladimirsx <Gallery /> ## Model description vladimirsx lora ## Trigger words You should use `vladimirsx` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/vladimirssx/tree/main) them in the Files & versions tab.
ericnunes1/phi4-r1
ericnunes1
2025-03-25T17:56:52Z
0
0
transformers
[ "transformers", "safetensors", "gguf", "llama", "text-generation", "conversational", "arxiv:1910.09700", "base_model:unsloth/phi-4-bnb-4bit", "base_model:quantized:unsloth/phi-4-bnb-4bit", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-25T16:58:21Z
--- base_model: unsloth/phi-4-bnb-4bit library_name: transformers license: apache-2.0 --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.14.0
aljebra/speecht5_tts_nigerian_accent
aljebra
2025-03-25T17:55:51Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "speecht5", "text-to-audio", "african-english, nigerian-accent, low-resource-tts", "generated_from_trainer", "en", "dataset:custom", "base_model:microsoft/speecht5_tts", "base_model:finetune:microsoft/speecht5_tts", "license:mit", "endpoints_compatible", "region:us" ]
text-to-audio
2025-03-25T08:35:31Z
--- library_name: transformers language: - en license: mit base_model: microsoft/speecht5_tts tags: - african-english, nigerian-accent, low-resource-tts - generated_from_trainer datasets: - custom model-index: - name: SpeechT5 TTS Nigerian English results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # SpeechT5 TTS Nigerian English This model is a fine-tuned version of [microsoft/speecht5_tts](https://huggingface.co/microsoft/speecht5_tts) on the Nigerian English TTS dataset. It achieves the following results on the evaluation set: - Loss: 0.4957 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 2 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 8 - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - training_steps: 4000 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-------:|:----:|:---------------:| | 0.5928 | 17.5487 | 1000 | 0.5506 | | 0.4771 | 35.0885 | 2000 | 0.4972 | | 0.4376 | 52.6372 | 3000 | 0.4875 | | 0.4249 | 70.1770 | 4000 | 0.4957 | ### Framework versions - Transformers 4.51.0.dev0 - Pytorch 2.6.0+cu124 - Datasets 3.4.1 - Tokenizers 0.21.1
kamel-usp/jbcs2025_bertimbau-large-C4
kamel-usp
2025-03-25T17:55:48Z
2
0
null
[ "safetensors", "bert", "aes", "pt", "en", "dataset:kamel-usp/aes_enem_dataset", "base_model:neuralmind/bert-large-portuguese-cased", "base_model:finetune:neuralmind/bert-large-portuguese-cased", "model-index", "region:us" ]
null
2025-03-16T00:57:28Z
--- language: - pt - en tags: - aes datasets: - kamel-usp/aes_enem_dataset base_model: neuralmind/bert-large-portuguese-cased metrics: - accuracy - qwk model-index: - name: bertimbau-large-C4 results: - task: type: text-classification name: Automated Essay Score dataset: name: Automated Essay Score ENEM Dataset type: kamel-usp/aes_enem_dataset config: JBCS2025 split: test metrics: - name: Macro F1 type: f1 value: 0.2947722798786629 - name: QWK type: qwk value: 0.5686184812442818 - name: Weighted Macro F1 type: f1 value: 0.554485898611708 --- # Model ID: bertimbau-large-C4 ## Results | | test_data | |:-----------------|------------:| | eval_accuracy | 0.528986 | | eval_RMSE | 30.8338 | | eval_QWK | 0.568618 | | eval_Macro_F1 | 0.294772 | | eval_Weighted_F1 | 0.554486 | | eval_Micro_F1 | 0.528986 | | eval_HDIV | 0.00724638 |
pictgensupport/beerv2
pictgensupport
2025-03-25T17:55:42Z
0
0
diffusers
[ "diffusers", "flux", "lora", "replicate", "text-to-image", "en", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "license:other", "region:us" ]
text-to-image
2025-03-25T17:55:40Z
--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md language: - en tags: - flux - diffusers - lora - replicate base_model: "black-forest-labs/FLUX.1-dev" pipeline_tag: text-to-image # widget: # - text: >- # prompt # output: # url: https://... instance_prompt: ICON_BASIC --- # Beerv2 <Gallery /> Trained on Replicate using: https://replicate.com/ostris/flux-dev-lora-trainer/train ## Trigger words You should use `ICON_BASIC` to trigger the image generation. ## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers) ```py from diffusers import AutoPipelineForText2Image import torch pipeline = AutoPipelineForText2Image.from_pretrained('black-forest-labs/FLUX.1-dev', torch_dtype=torch.float16).to('cuda') pipeline.load_lora_weights('pictgensupport/beerv2', weight_name='lora.safetensors') image = pipeline('your prompt').images[0] ``` For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters)
LHRuig/lagertsx
LHRuig
2025-03-25T17:54:24Z
0
0
diffusers
[ "diffusers", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T17:53:08Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: lagertsx --- # lagertsx <Gallery /> ## Model description lagertsx lora ## Trigger words You should use `lagertsx` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/lagertsx/tree/main) them in the Files & versions tab.
Alphatao/a6f4ed88-0806-4935-9632-dbde555fae5c
Alphatao
2025-03-25T17:54:07Z
0
0
peft
[ "peft", "safetensors", "llama", "axolotl", "generated_from_trainer", "base_model:MLP-KTLim/llama-3-Korean-Bllossom-8B", "base_model:adapter:MLP-KTLim/llama-3-Korean-Bllossom-8B", "license:llama3", "region:us" ]
null
2025-03-25T11:32:20Z
--- library_name: peft license: llama3 base_model: MLP-KTLim/llama-3-Korean-Bllossom-8B tags: - axolotl - generated_from_trainer model-index: - name: a6f4ed88-0806-4935-9632-dbde555fae5c results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.4.1` ```yaml adapter: lora base_model: MLP-KTLim/llama-3-Korean-Bllossom-8B bf16: true chat_template: llama3 dataset_prepared_path: null datasets: - data_files: - 614113b4f1a6b045_train_data.json ds_type: json format: custom path: /workspace/input_data/614113b4f1a6b045_train_data.json type: field_input: examples field_instruction: func_desc field_output: answers format: '{instruction} {input}' no_input_format: '{instruction}' system_format: '{system}' system_prompt: '' debug: null device_map: ? '' : 0,1,2,3,4,5,6,7 early_stopping_patience: 2 eval_max_new_tokens: 128 eval_steps: 400 eval_table_size: null flash_attention: true gradient_accumulation_steps: 4 gradient_checkpointing: true group_by_length: false hub_model_id: Alphatao/a6f4ed88-0806-4935-9632-dbde555fae5c hub_repo: null hub_strategy: null hub_token: null learning_rate: 0.0002 load_best_model_at_end: true load_in_4bit: false load_in_8bit: false local_rank: null logging_steps: 1 lora_alpha: 32 lora_dropout: 0.05 lora_fan_in_fan_out: null lora_model_dir: null lora_r: 16 lora_target_linear: true lora_target_modules: - q_proj - k_proj - v_proj - o_proj - down_proj - up_proj lr_scheduler: cosine max_grad_norm: 1.0 max_steps: 4875 micro_batch_size: 2 mlflow_experiment_name: /tmp/614113b4f1a6b045_train_data.json model_type: AutoModelForCausalLM num_epochs: 2 optimizer: adamw_bnb_8bit output_dir: miner_id_24 pad_to_sequence_len: true resume_from_checkpoint: null s2_attention: null sample_packing: false save_steps: 400 sequence_len: 2048 strict: false tf32: true tokenizer_type: AutoTokenizer train_on_inputs: false trust_remote_code: true val_set_size: 0.04 wandb_entity: null wandb_mode: online wandb_name: 923afc99-aa10-4ac7-925a-f5275d76ccd4 wandb_project: Gradients-On-Demand wandb_run: your_name wandb_runid: 923afc99-aa10-4ac7-925a-f5275d76ccd4 warmup_steps: 10 weight_decay: 0.0 xformers_attention: null ``` </details><br> # a6f4ed88-0806-4935-9632-dbde555fae5c This model is a fine-tuned version of [MLP-KTLim/llama-3-Korean-Bllossom-8B](https://huggingface.co/MLP-KTLim/llama-3-Korean-Bllossom-8B) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.4071 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 2 - eval_batch_size: 2 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 8 - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 10 - training_steps: 4875 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:------:|:----:|:---------------:| | 1.3985 | 0.0002 | 1 | 2.2834 | | 0.4249 | 0.0850 | 400 | 0.4947 | | 0.7992 | 0.1700 | 800 | 0.4869 | | 0.3926 | 0.2549 | 1200 | 0.4829 | | 0.5273 | 0.3399 | 1600 | 0.4761 | | 0.4743 | 0.4249 | 2000 | 0.4581 | | 0.405 | 0.5099 | 2400 | 0.4388 | | 0.4932 | 0.5948 | 2800 | 0.4350 | | 0.3592 | 0.6798 | 3200 | 0.4227 | | 0.3654 | 0.7648 | 3600 | 0.4171 | | 0.3903 | 0.8498 | 4000 | 0.4110 | | 0.509 | 0.9347 | 4400 | 0.4060 | | 0.4469 | 1.0197 | 4800 | 0.4071 | ### Framework versions - PEFT 0.13.2 - Transformers 4.46.0 - Pytorch 2.5.0+cu124 - Datasets 3.0.1 - Tokenizers 0.20.1
LHRuig/chinesebysx
LHRuig
2025-03-25T17:52:39Z
0
0
diffusers
[ "diffusers", "safetensors", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T17:52:31Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: chinesebysx --- # chinesebysx <Gallery /> ## Model description chinesebysx lora ## Trigger words You should use `chinesebysx` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/chinesebysx/tree/main) them in the Files & versions tab.
LHRuig/vladmirmsx
LHRuig
2025-03-25T17:52:09Z
0
0
diffusers
[ "diffusers", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T17:51:59Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: vladmirmsx --- # vladmirmsx <Gallery /> ## Model description vladmirmsx lora ## Trigger words You should use `vladmirmsx` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/vladmirmsx/tree/main) them in the Files & versions tab.
jiggyjo11/qwen2vl_2b_aubrey
jiggyjo11
2025-03-25T17:51:41Z
0
0
transformers
[ "transformers", "safetensors", "qwen2_vl", "image-text-to-text", "text-generation", "conversational", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-25T17:26:37Z
--- library_name: transformers pipeline_tag: text-generation ---
LHRuig/rayzhangsx
LHRuig
2025-03-25T17:51:28Z
0
0
diffusers
[ "diffusers", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T17:51:22Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: rayzhangsx --- # rayzhangsx <Gallery /> ## Model description rayzhangsx lora ## Trigger words You should use `rayzhangsx` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/rayzhangsx/tree/main) them in the Files & versions tab.
kamel-usp/jbcs2025_bertimbau-large-C3
kamel-usp
2025-03-25T17:51:20Z
2
0
null
[ "safetensors", "bert", "aes", "pt", "en", "dataset:kamel-usp/aes_enem_dataset", "base_model:neuralmind/bert-large-portuguese-cased", "base_model:finetune:neuralmind/bert-large-portuguese-cased", "model-index", "region:us" ]
null
2025-03-16T00:40:54Z
--- language: - pt - en tags: - aes datasets: - kamel-usp/aes_enem_dataset base_model: neuralmind/bert-large-portuguese-cased metrics: - accuracy - qwk model-index: - name: bertimbau-large-C3 results: - task: type: text-classification name: Automated Essay Score dataset: name: Automated Essay Score ENEM Dataset type: kamel-usp/aes_enem_dataset config: JBCS2025 split: test metrics: - name: Macro F1 type: f1 value: 0.19411606228274925 - name: QWK type: qwk value: 0.26937738246505727 - name: Weighted Macro F1 type: f1 value: 0.24825898925023224 --- # Model ID: bertimbau-large-C3 ## Results | | test_data | |:-----------------|------------:| | eval_accuracy | 0.289855 | | eval_RMSE | 51.0754 | | eval_QWK | 0.269377 | | eval_Macro_F1 | 0.194116 | | eval_Weighted_F1 | 0.248259 | | eval_Micro_F1 | 0.289855 | | eval_HDIV | 0.0217391 |
LHRuig/cineanamorphicsx
LHRuig
2025-03-25T17:51:09Z
0
0
diffusers
[ "diffusers", "text-to-image", "lora", "template:diffusion-lora", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "region:us" ]
text-to-image
2025-03-25T17:50:29Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: suit output: url: images/suit.jpg base_model: black-forest-labs/FLUX.1-dev instance_prompt: man --- # cineanamorphicsx <Gallery /> ## Model description cineanamorphicsx lora ## Trigger words You should use `man` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/LHRuig/cineanamorphicsx/tree/main) them in the Files & versions tab.
brothersen/food-classifier
brothersen
2025-03-25T17:49:14Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "vit", "image-classification", "generated_from_trainer", "base_model:google/vit-base-patch16-224-in21k", "base_model:finetune:google/vit-base-patch16-224-in21k", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
image-classification
2025-03-25T16:54:49Z
--- library_name: transformers license: apache-2.0 base_model: google/vit-base-patch16-224-in21k tags: - generated_from_trainer metrics: - accuracy model-index: - name: food-classifier results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # food-classifier This model is a fine-tuned version of [google/vit-base-patch16-224-in21k](https://huggingface.co/google/vit-base-patch16-224-in21k) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 1.6384 - Accuracy: 0.892 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 64 - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - num_epochs: 3 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 2.5596 | 1.0 | 63 | 2.4049 | 0.837 | | 1.871 | 2.0 | 126 | 1.7607 | 0.895 | | 1.6474 | 2.96 | 186 | 1.6384 | 0.892 | ### Framework versions - Transformers 4.48.3 - Pytorch 2.6.0+cpu - Datasets 2.16.1 - Tokenizers 0.21.0
shrey123354/lesia_all_auto
shrey123354
2025-03-25T17:48:29Z
0
0
diffusers
[ "diffusers", "flux", "lora", "replicate", "text-to-image", "en", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "license:other", "region:us" ]
text-to-image
2025-03-25T17:06:39Z
--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md language: - en tags: - flux - diffusers - lora - replicate base_model: "black-forest-labs/FLUX.1-dev" pipeline_tag: text-to-image # widget: # - text: >- # prompt # output: # url: https://... instance_prompt: Sidf --- # Lesia_All_Auto <Gallery /> Trained on Replicate using: https://replicate.com/ostris/flux-dev-lora-trainer/train ## Trigger words You should use `Sidf` to trigger the image generation. ## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers) ```py from diffusers import AutoPipelineForText2Image import torch pipeline = AutoPipelineForText2Image.from_pretrained('black-forest-labs/FLUX.1-dev', torch_dtype=torch.float16).to('cuda') pipeline.load_lora_weights('shrey123354/lesia_all_auto', weight_name='lora.safetensors') image = pipeline('your prompt').images[0] ``` For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters)
Tapendra/gemma-3-4b-it_checkpoint_v6
Tapendra
2025-03-25T17:48:08Z
0
0
peft
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:google/gemma-3-4b-it", "base_model:adapter:google/gemma-3-4b-it", "region:us" ]
null
2025-03-25T17:47:40Z
--- base_model: google/gemma-3-4b-it library_name: peft --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.14.0
shettysach/dsq_1.5b_fc
shettysach
2025-03-25T17:46:45Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "generated_from_trainer", "trl", "sft", "base_model:deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B", "base_model:finetune:deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B", "endpoints_compatible", "region:us" ]
null
2025-03-25T04:33:21Z
--- base_model: deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B library_name: transformers model_name: dsq_1.5b_fc tags: - generated_from_trainer - trl - sft licence: license --- # Model Card for dsq_1.5b_fc This model is a fine-tuned version of [deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B). It has been trained using [TRL](https://github.com/huggingface/trl). ## Quick start ```python from transformers import pipeline question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?" generator = pipeline("text-generation", model="shettysach/dsq_1.5b_fc", device="cuda") output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0] print(output["generated_text"]) ``` ## Training procedure This model was trained with SFT. ### Framework versions - TRL: 0.16.0 - Transformers: 4.50.0 - Pytorch: 2.6.0+cu124 - Datasets: 3.4.1 - Tokenizers: 0.21.1 ## Citations Cite TRL as: ```bibtex @misc{vonwerra2022trl, title = {{TRL: Transformer Reinforcement Learning}}, author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec}, year = 2020, journal = {GitHub repository}, publisher = {GitHub}, howpublished = {\url{https://github.com/huggingface/trl}} } ```
genki10/BERT_AugV8_k5_task1_organization_sp040_lw030_fold3
genki10
2025-03-25T17:46:42Z
0
0
transformers
[ "transformers", "pytorch", "bert", "text-classification", "generated_from_trainer", "base_model:google-bert/bert-base-uncased", "base_model:finetune:google-bert/bert-base-uncased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2025-03-25T17:36:27Z
--- library_name: transformers license: apache-2.0 base_model: bert-base-uncased tags: - generated_from_trainer model-index: - name: BERT_AugV8_k5_task1_organization_sp040_lw030_fold3 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # BERT_AugV8_k5_task1_organization_sp040_lw030_fold3 This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.2649 - Qwk: 0.2853 - Mse: 1.2652 - Rmse: 1.1248 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - num_epochs: 150 ### Training results | Training Loss | Epoch | Step | Validation Loss | Qwk | Mse | Rmse | |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|:------:| | No log | 1.0 | 4 | 8.8585 | 0.0 | 8.8567 | 2.9760 | | No log | 2.0 | 8 | 5.6138 | 0.0210 | 5.6124 | 2.3691 | | No log | 3.0 | 12 | 3.6073 | 0.0 | 3.6060 | 1.8990 | | No log | 4.0 | 16 | 2.1295 | 0.1042 | 2.1287 | 1.4590 | | No log | 5.0 | 20 | 1.4222 | 0.0365 | 1.4215 | 1.1923 | | No log | 6.0 | 24 | 1.0952 | 0.0202 | 1.0945 | 1.0462 | | No log | 7.0 | 28 | 2.2531 | 0.0425 | 2.2522 | 1.5007 | | No log | 8.0 | 32 | 1.0485 | 0.0935 | 1.0479 | 1.0237 | | No log | 9.0 | 36 | 1.5476 | 0.1079 | 1.5470 | 1.2438 | | No log | 10.0 | 40 | 0.9696 | 0.2235 | 0.9694 | 0.9846 | | No log | 11.0 | 44 | 1.1783 | 0.2603 | 1.1787 | 1.0857 | | No log | 12.0 | 48 | 0.7040 | 0.4612 | 0.7046 | 0.8394 | | No log | 13.0 | 52 | 0.9743 | 0.3493 | 0.9747 | 0.9873 | | No log | 14.0 | 56 | 0.9616 | 0.3546 | 0.9618 | 0.9807 | | No log | 15.0 | 60 | 0.8223 | 0.4309 | 0.8228 | 0.9071 | | No log | 16.0 | 64 | 1.3972 | 0.3270 | 1.3977 | 1.1823 | | No log | 17.0 | 68 | 1.0623 | 0.3650 | 1.0631 | 1.0311 | | No log | 18.0 | 72 | 1.0269 | 0.3507 | 1.0276 | 1.0137 | | No log | 19.0 | 76 | 1.8963 | 0.1850 | 1.8968 | 1.3773 | | No log | 20.0 | 80 | 1.0329 | 0.3345 | 1.0336 | 1.0167 | | No log | 21.0 | 84 | 1.0305 | 0.3449 | 1.0312 | 1.0155 | | No log | 22.0 | 88 | 1.8547 | 0.1776 | 1.8550 | 1.3620 | | No log | 23.0 | 92 | 0.8733 | 0.3422 | 0.8737 | 0.9347 | | No log | 24.0 | 96 | 1.7834 | 0.1884 | 1.7838 | 1.3356 | | No log | 25.0 | 100 | 1.4348 | 0.2453 | 1.4350 | 1.1979 | | No log | 26.0 | 104 | 1.0677 | 0.3069 | 1.0680 | 1.0334 | | No log | 27.0 | 108 | 1.2649 | 0.2853 | 1.2652 | 1.1248 | ### Framework versions - Transformers 4.47.0 - Pytorch 2.5.1+cu121 - Datasets 3.3.1 - Tokenizers 0.21.0
pictgensupport/basketv2
pictgensupport
2025-03-25T17:46:21Z
0
0
diffusers
[ "diffusers", "flux", "lora", "replicate", "text-to-image", "en", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "license:other", "region:us" ]
text-to-image
2025-03-25T17:46:19Z
--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md language: - en tags: - flux - diffusers - lora - replicate base_model: "black-forest-labs/FLUX.1-dev" pipeline_tag: text-to-image # widget: # - text: >- # prompt # output: # url: https://... instance_prompt: ICON_BASIC --- # Basketv2 <Gallery /> Trained on Replicate using: https://replicate.com/ostris/flux-dev-lora-trainer/train ## Trigger words You should use `ICON_BASIC` to trigger the image generation. ## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers) ```py from diffusers import AutoPipelineForText2Image import torch pipeline = AutoPipelineForText2Image.from_pretrained('black-forest-labs/FLUX.1-dev', torch_dtype=torch.float16).to('cuda') pipeline.load_lora_weights('pictgensupport/basketv2', weight_name='lora.safetensors') image = pipeline('your prompt').images[0] ``` For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters)
ethansandbar/3_25_2025_6
ethansandbar
2025-03-25T17:46:10Z
0
0
transformers
[ "transformers", "safetensors", "generated_from_trainer", "trl", "sft", "base_model:google/gemma-3-12b-it", "base_model:finetune:google/gemma-3-12b-it", "endpoints_compatible", "region:us" ]
null
2025-03-25T17:46:06Z
--- base_model: google/gemma-3-12b-it library_name: transformers model_name: '3_25_2025_6' tags: - generated_from_trainer - trl - sft licence: license --- # Model Card for 3_25_2025_6 This model is a fine-tuned version of [google/gemma-3-12b-it](https://huggingface.co/google/gemma-3-12b-it). It has been trained using [TRL](https://github.com/huggingface/trl). ## Quick start ```python from transformers import pipeline question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?" generator = pipeline("text-generation", model="ethansandbar/3_25_2025_6", device="cuda") output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0] print(output["generated_text"]) ``` ## Training procedure This model was trained with SFT. ### Framework versions - TRL: 0.15.2 - Transformers: 4.50.0.dev0 - Pytorch: 2.6.0 - Datasets: 3.4.1 - Tokenizers: 0.21.1 ## Citations Cite TRL as: ```bibtex @misc{vonwerra2022trl, title = {{TRL: Transformer Reinforcement Learning}}, author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec}, year = 2020, journal = {GitHub repository}, publisher = {GitHub}, howpublished = {\url{https://github.com/huggingface/trl}} } ```
lesso10/95cb9f7a-8299-4794-a4b0-7bb4e17ae931
lesso10
2025-03-25T17:45:14Z
0
0
peft
[ "peft", "safetensors", "phi", "axolotl", "generated_from_trainer", "base_model:echarlaix/tiny-random-PhiForCausalLM", "base_model:adapter:echarlaix/tiny-random-PhiForCausalLM", "license:apache-2.0", "region:us" ]
null
2025-03-25T15:42:36Z
--- library_name: peft license: apache-2.0 base_model: echarlaix/tiny-random-PhiForCausalLM tags: - axolotl - generated_from_trainer model-index: - name: 95cb9f7a-8299-4794-a4b0-7bb4e17ae931 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.4.1` ```yaml adapter: lora base_model: echarlaix/tiny-random-PhiForCausalLM bf16: auto chat_template: llama3 dataset_prepared_path: null datasets: - data_files: - 7462b07f6259b24d_train_data.json ds_type: json format: custom path: /workspace/input_data/7462b07f6259b24d_train_data.json type: field_instruction: startphrase field_output: gold-ending format: '{instruction}' no_input_format: '{instruction}' system_format: '{system}' system_prompt: '' debug: null deepspeed: null do_eval: true early_stopping_patience: 3 eval_batch_size: 4 eval_max_new_tokens: 128 eval_steps: 500 evals_per_epoch: null flash_attention: true fp16: false fsdp: null fsdp_config: null gradient_accumulation_steps: 8 gradient_checkpointing: true group_by_length: true hub_model_id: lesso10/95cb9f7a-8299-4794-a4b0-7bb4e17ae931 hub_repo: null hub_strategy: checkpoint hub_token: null learning_rate: 0.00021 load_in_4bit: false load_in_8bit: false local_rank: null logging_steps: 50 lora_alpha: 128 lora_dropout: 0.15 lora_fan_in_fan_out: null lora_model_dir: null lora_r: 64 lora_target_linear: true lr_scheduler: cosine max_grad_norm: 1.0 max_steps: 50000 micro_batch_size: 4 mlflow_experiment_name: /tmp/7462b07f6259b24d_train_data.json model_type: AutoModelForCausalLM num_epochs: 10 optimizer: adamw_torch_fused output_dir: miner_id_24 pad_to_sequence_len: true resume_from_checkpoint: null s2_attention: null sample_packing: false save_steps: 500 saves_per_epoch: null seed: 100 sequence_len: 1024 special_tokens: pad_token: <|endoftext|> strict: false tf32: true tokenizer_type: AutoTokenizer train_on_inputs: false trust_remote_code: true val_set_size: 0.05 wandb_entity: null wandb_mode: online wandb_name: 9611c628-3f80-4127-8fd5-47e5a88912ed wandb_project: 10a wandb_run: your_name wandb_runid: 9611c628-3f80-4127-8fd5-47e5a88912ed warmup_steps: 100 weight_decay: 0.0 xformers_attention: null ``` </details><br> # 95cb9f7a-8299-4794-a4b0-7bb4e17ae931 This model is a fine-tuned version of [echarlaix/tiny-random-PhiForCausalLM](https://huggingface.co/echarlaix/tiny-random-PhiForCausalLM) on the None dataset. It achieves the following results on the evaluation set: - Loss: 6.7384 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.00021 - train_batch_size: 4 - eval_batch_size: 4 - seed: 100 - gradient_accumulation_steps: 8 - total_train_batch_size: 32 - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 100 - training_steps: 27536 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:------:|:-----:|:---------------:| | No log | 0.0004 | 1 | 6.9352 | | 6.8024 | 0.1816 | 500 | 6.7963 | | 6.7901 | 0.3632 | 1000 | 6.7815 | | 6.7796 | 0.5447 | 1500 | 6.7703 | | 6.779 | 0.7263 | 2000 | 6.7642 | | 6.7745 | 0.9079 | 2500 | 6.7608 | | 6.7737 | 1.0896 | 3000 | 6.7579 | | 6.7684 | 1.2712 | 3500 | 6.7563 | | 6.7663 | 1.4528 | 4000 | 6.7536 | | 6.7712 | 1.6343 | 4500 | 6.7521 | | 6.7621 | 1.8159 | 5000 | 6.7513 | | 6.767 | 1.9975 | 5500 | 6.7501 | | 6.7656 | 2.1792 | 6000 | 6.7488 | | 6.7617 | 2.3608 | 6500 | 6.7480 | | 6.767 | 2.5424 | 7000 | 6.7474 | | 6.7633 | 2.7240 | 7500 | 6.7459 | | 6.7622 | 2.9055 | 8000 | 6.7454 | | 6.7648 | 3.0872 | 8500 | 6.7448 | | 6.7651 | 3.2688 | 9000 | 6.7445 | | 6.7563 | 3.4504 | 9500 | 6.7443 | | 6.7515 | 3.6320 | 10000 | 6.7435 | | 6.7584 | 3.8136 | 10500 | 6.7433 | | 6.7575 | 3.9951 | 11000 | 6.7429 | | 6.7601 | 4.1769 | 11500 | 6.7421 | | 6.7569 | 4.3584 | 12000 | 6.7421 | | 6.7588 | 4.5400 | 12500 | 6.7422 | | 6.7567 | 4.7216 | 13000 | 6.7418 | | 6.7589 | 4.9032 | 13500 | 6.7416 | | 6.7552 | 5.0849 | 14000 | 6.7409 | | 6.7603 | 5.2665 | 14500 | 6.7410 | | 6.7528 | 5.4480 | 15000 | 6.7406 | | 6.761 | 5.6296 | 15500 | 6.7404 | | 6.7526 | 5.8112 | 16000 | 6.7400 | | 6.758 | 5.9928 | 16500 | 6.7401 | | 6.762 | 6.1745 | 17000 | 6.7399 | | 6.7445 | 6.3561 | 17500 | 6.7395 | | 6.7551 | 6.5377 | 18000 | 6.7395 | | 6.755 | 6.7192 | 18500 | 6.7395 | | 6.7563 | 6.9008 | 19000 | 6.7389 | | 6.7548 | 7.0825 | 19500 | 6.7390 | | 6.7522 | 7.2641 | 20000 | 6.7390 | | 6.7529 | 7.4457 | 20500 | 6.7386 | | 6.7504 | 7.6273 | 21000 | 6.7385 | | 6.7464 | 7.8088 | 21500 | 6.7386 | | 6.7592 | 7.9904 | 22000 | 6.7385 | | 6.7546 | 8.1721 | 22500 | 6.7385 | | 6.7543 | 8.3537 | 23000 | 6.7383 | | 6.7548 | 8.5353 | 23500 | 6.7385 | | 6.7547 | 8.7169 | 24000 | 6.7383 | | 6.7531 | 8.8985 | 24500 | 6.7381 | | 6.757 | 9.0802 | 25000 | 6.7384 | | 6.751 | 9.2617 | 25500 | 6.7381 | | 6.7435 | 9.4433 | 26000 | 6.7383 | | 6.7603 | 9.6249 | 26500 | 6.7382 | | 6.7518 | 9.8065 | 27000 | 6.7384 | ### Framework versions - PEFT 0.13.2 - Transformers 4.46.0 - Pytorch 2.5.0+cu124 - Datasets 3.0.1 - Tokenizers 0.20.1
gradientrouting-spar/ft_qwen_v3_seed2_1dproxy
gradientrouting-spar
2025-03-25T17:44:55Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2025-03-25T17:44:28Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
eming/iLoRA
eming
2025-03-25T17:44:30Z
0
0
null
[ "region:us" ]
null
2025-03-25T16:56:16Z
# iLoRA #### Preparation 1. Prepare the environment: ```python git clone cd iLoRA pip install -r requirements.txt ``` 2. Prepare the pre-trained huggingface model of Llama2-7B (https://huggingface.co/meta-llama/Llama-2-7b-hf). 3. Modify the paths inside the .sh file. #### Train iLoRA Train iLoRA with a single A100 GPU on MovieLens dataset: ```python sh train_movielens.sh ``` Train iLoRA with a single A100 GPU on Steam dataset: ``` sh train_steam.sh ``` Train iLoRA with a single A100 GPU on LastFM dataset: ``` sh train_lastfm.sh ``` Note that: set the `llm_path` argument with your own directory path of the Llama2 model. ##### For the environmental issues mentioned by everyone during the reproduction process, we have attempted to help resolve them and have listed some solutions: If you encounter an error with your transformers/generation/utils.py file, please replace it with the debug/utils.py file we have provided in your environment. If you encounter an error with your transformers/models/llama/modeling_llama.py file, please replace it with the debug/modeling_llama.py file. Thank you all for your attention to our work! Wishing you success in your research. ##### Evaluate iLoRA Test iLoRA with a single A100 GPU on MovieLens dataset: ``` sh test_movielens.sh ``` Test iLoRA with a single A100 GPU on Steam dataset: ``` sh test_steam.sh ``` Test iLoRA with a single A100 GPU on LastFM dataset: ``` sh test_lastfm.sh ```
FluxiIA/Translate_ENPT_PTEN-GRPO
FluxiIA
2025-03-25T17:43:58Z
0
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "text-generation-inference", "unsloth", "trl", "grpo", "conversational", "en", "base_model:FluxiIA/translate2_kto_full", "base_model:finetune:FluxiIA/translate2_kto_full", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2025-03-25T17:38:45Z
--- base_model: FluxiIA/translate2_kto_full tags: - text-generation-inference - transformers - unsloth - llama - trl - grpo license: apache-2.0 language: - en --- # Uploaded model - **Developed by:** FluxiIA - **License:** apache-2.0 - **Finetuned from model :** FluxiIA/translate2_kto_full This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth) ```json EN_TO_PT_PROMPT = """ Traduza o texto de EN PARA PT """ PT_TO_EN_PROMPT = """ Traduza o texto de PT PARA EN """ ```
Metaskepsis/W
Metaskepsis
2025-03-25T17:43:14Z
0
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-25T17:33:51Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
wetey/MARBERT-LHSAB
wetey
2025-03-25T17:43:13Z
0
0
transformers
[ "transformers", "safetensors", "bert", "text-classification", "offensive language detection", "ar", "base_model:UBC-NLP/MARBERT", "base_model:finetune:UBC-NLP/MARBERT", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-07-15T19:34:54Z
--- license: mit language: - ar metrics: - accuracy - f1 - precision - recall library_name: transformers tags: - offensive language detection base_model: - UBC-NLP/MARBERT --- This model is part of the work done in <!-- add paper name -->. <br> The full code can be found at https://github.com/wetey/cluster-errors ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Model type:** BERT-based - **Language(s) (NLP):** Arabic - **Finetuned from model:** UBC-NLP/MARBERT ## How to Get Started with the Model Use the code below to get started with the model. ```python # Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-classification", model="wetey/MARBERT-LHSAB") ``` ```python # Load model directly from transformers import AutoTokenizer, AutoModelForSequenceClassification tokenizer = AutoTokenizer.from_pretrained("wetey/MARBERT-LHSAB") model = AutoModelForSequenceClassification.from_pretrained("wetey/MARBERT-LHSAB") ``` ## Fine-tuning Details ### Fine-tuning Data This model is fine-tuned on the [L-HSAB](https://github.com/Hala-Mulki/L-HSAB-First-Arabic-Levantine-HateSpeech-Dataset). The exact version we use (after removing duplicates) can be found [](). <!--TODO--> ### Fine-tuning Procedure The exact fine-tuning procedure followed can be found [here](https://github.com/wetey/cluster-errors/tree/master/finetuning) #### Training Hyperparameters evaluation_strategy = 'epoch' logging_steps = 1, num_train_epochs = 5, learning_rate = 1e-5, eval_accumulation_steps = 2 ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data Test set used can be found [here](https://github.com/wetey/cluster-errors/tree/master/data/datasets) ### Results `accuracy`: 87.9% <br> `precision`: 88.1% <br> `recall`: 87.9% <br> `f1-score`: 87.9% <br> #### Results per class | Label | Precision | Recall | F1-score| |---------|---------|---------|---------| | normal | 85% | 82% | 83% | | abusive | 93% | 92% | 93% | | hate | 68% | 78% | 72% | ## Citation <!--TODO-->
martian786/fined_tuned_all-mpnet-base-trec_clinical_trials
martian786
2025-03-25T17:43:08Z
0
0
null
[ "safetensors", "mpnet", "dataset:irds/clinicaltrials_2021_trec-ct-2021", "base_model:sentence-transformers/all-mpnet-base-v2", "base_model:finetune:sentence-transformers/all-mpnet-base-v2", "license:apache-2.0", "region:us" ]
null
2025-03-25T14:45:08Z
--- license: apache-2.0 base_model: - sentence-transformers/all-mpnet-base-v2 metrics: - precision - accuracy datasets: - irds/clinicaltrials_2021_trec-ct-2021 --- This model is one that has been fined tuned on TREC 2021 Clinical Trials data. The base model used is all-mpnet-base-v2 Model Description This model is a fine‑tuned version of the sentence-transformers/all-mpnet-base-v2 model. It has been specialized for the clinical trials domain using the TREC-CT 2021 dataset. The fine‑tuning process utilized a triplet loss approach on training triplets created from: Queries: Natural language queries from the clinical trial retrieval task. Documents: Clinical trial documents (using fields like detailed_descrption or summary). Relevance Judgments (qrels): Labelled data indicating document relevance. This model produces sentence embeddings that capture subtle clinical nuances, making it well‑suited for semantic search and retrieval applications in the clinical trials space. . Intended Uses Semantic Search & Retrieval: Retrieve relevant clinical trial documents based on user queries. Information Extraction: Enhance retrieval pipelines in clinical research platforms. Academic Research: Serve as a baseline for further research in clinical trials information retrieval. How to Use You can load and use the model with the Hugging Face Transformers library: from transformers import AutoTokenizer, AutoModel # Load the fine-tuned model and tokenizer tokenizer = AutoTokenizer.from_pretrained("martian786/fined_tuned_all-mpnet-base-trec_clinical_trials") model = AutoModel.from_pretrained("martian786/fined_tuned_all-mpnet-base-trec_clinical_trials") # Example: Get embedding for a clinical trial query def mean_pooling(model_output, attention_mask): token_embeddings = model_output[0] # First element contains token embeddings input_mask_expanded = attention_mask.unsqueeze(-1).expand(token_embeddings.size()).float() return torch.sum(token_embeddings * input_mask_expanded, 1) / torch.clamp(input_mask_expanded.sum(1), min=1e-9) def get_embedding(text): encoded_input = tokenizer(text, padding="max_length", truncation=True, max_length=512, return_tensors="pt") with torch.no_grad(): model_output = model(**encoded_input) embedding = mean_pooling(model_output, encoded_input['attention_mask']) return embedding[0].cpu().numpy().tolist() Training Data The model was fine‑tuned using the TREC-CT 2021 dataset, which comprises: Documents: Clinical trial documents with fields like doc_id, title, summary, detailed_descrption, and eligibility. Queries: Clinical queries provided in trec_ct_2021_query.jsonl. Relevance Judgments (qrels): Judgements from trec_ct_2021_qrels.jsonl used to generate training triplets (anchor, positive, negative). Evaluation The fine‑tuning process was designed to improve retrieval performance measured by metrics such as: Precision@K Recall@K F1 Score Average Precision (AP) nDCG MRR Evaluation experiments demonstrated improved retrieval performance on the TREC-CT 2021 clinical trials data compared to the base model. Limitations Domain Specificity: The model is specialized for clinical trials retrieval based on the TREC-CT 2021 dataset. Its performance on other clinical or non‑clinical datasets may vary. Data Bias: As with any model trained on a specific dataset, biases present in the TREC-CT 2021 data may influence retrieval performance. Query Variability: Very verbose or non-standard queries might still challenge the model. Additional fine‑tuning or query preprocessing might be necessary in some scenarios.
kamel-usp/jbcs2025_mbert_base-C4
kamel-usp
2025-03-25T17:42:57Z
5
0
null
[ "safetensors", "bert", "aes", "pt", "en", "dataset:kamel-usp/aes_enem_dataset", "base_model:google-bert/bert-base-multilingual-cased", "base_model:finetune:google-bert/bert-base-multilingual-cased", "model-index", "region:us" ]
null
2025-03-15T23:47:56Z
--- language: - pt - en tags: - aes datasets: - kamel-usp/aes_enem_dataset base_model: google-bert/bert-base-multilingual-cased metrics: - accuracy - qwk model-index: - name: mbert_base-C4 results: - task: type: text-classification name: Automated Essay Score dataset: name: Automated Essay Score ENEM Dataset type: kamel-usp/aes_enem_dataset config: JBCS2025 split: test metrics: - name: Macro F1 type: f1 value: 0.17299898682877404 - name: QWK type: qwk value: 0.28170809432759725 - name: Weighted Macro F1 type: f1 value: 0.4091229461257213 --- # Model ID: mbert_base-C4 ## Results | | test_data | |:-----------------|------------:| | eval_accuracy | 0.5 | | eval_RMSE | 33.708 | | eval_QWK | 0.281708 | | eval_Macro_F1 | 0.172999 | | eval_Weighted_F1 | 0.409123 | | eval_Micro_F1 | 0.5 | | eval_HDIV | 0.00724638 |
wetey/distilbert-base-uncased-measuring-hate-speech
wetey
2025-03-25T17:42:03Z
0
0
transformers
[ "transformers", "safetensors", "distilbert", "text-classification", "offensive language detection ", "en", "dataset:ucberkeley-dlab/measuring-hate-speech", "base_model:distilbert/distilbert-base-uncased", "base_model:finetune:distilbert/distilbert-base-uncased", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2024-07-15T19:31:02Z
--- license: mit datasets: - ucberkeley-dlab/measuring-hate-speech language: - en metrics: - accuracy - f1 - precision - recall library_name: transformers pipeline_tag: text-classification tags: - 'offensive language detection ' base_model: - distilbert/distilbert-base-uncased --- This model is part of the work done in <!-- add paper name -->. <br> The full code can be found at https://github.com/wetey/cluster-errors ## Model Details ### Model Description - **Model type:** Distil-BERT - **Language(s) (NLP):** English - **Finetuned from model:** distilbert-base-uncased ## How to Get Started with the Model Use the code below to get started with the model. ```python # Use a pipeline as a high-level helper from transformers import pipeline pipe = pipeline("text-classification", model="wetey/distilbert-base-uncased-measuring-hate-speech") ``` ```python # Load model directly from transformers import AutoTokenizer, AutoModelForSequenceClassification tokenizer = AutoTokenizer.from_pretrained("wetey/distilbert-base-uncased-measuring-hate-speech") model = AutoModelForSequenceClassification.from_pretrained("wetey/distilbert-base-uncased-measuring-hate-speech") ``` ## Fine-tuning Details ### Fine-tuning Data The model was fine-tuned on the [ucberkeley-dlab/measuring-hate-speech](https://huggingface.co/datasets/ucberkeley-dlab/measuring-hate-speech) dataset. <br> We converted the continuous hatespeech scores to categorical labels using the ranges suggested by the authors. The ranges are listed on the [HuggingFace Dataset card](https://huggingface.co/datasets/ucberkeley-dlab/measuring-hate-speech). <br> Examples with hatespeech scores that are lower than -1 are considered `supportive`, between -1 and 0.5 are `neutral`, and scores greater than 0.5 are `hatespeech`. <br> We remove duplicate examples along with those that received fewer than three total annotations, and we drop the neutral class. <br> After these steps, we were left with 12,289 examples with 7497 examples labeled as `supportive` and 4792 labeled as `hatespeech`. We use 85\% of the dataset for fine-tuning and 15\% for testing. ### Fine-tuning Procedure The exact fine-tuning procedure followed can be found [here](https://github.com/wetey/cluster-errors/tree/master/finetuning) #### Fine-tuning Hyperparameters evaluation_strategy = 'epoch' logging_steps = 1, num_train_epochs = 5, learning_rate = 1e-5, eval_accumulation_steps = 2 ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data Test set used can be found [here](https://github.com/wetey/cluster-errors/tree/master/data/datasets) ### Results `accuracy`: 89.3% <br> `precision`: 89.4% <br> `recall`: 89.3% <br> `f1-score`: 89.3% <br> #### Results per class | Label | Precision | Recall | F1-score| |---------|---------|---------|---------| | supportive | 92% | 91% | 91% | | hatespeech| 86% | 87% | 86% | ## Citation <!--TODO-->
EduCampozan/modelEduardo
EduCampozan
2025-03-25T17:41:21Z
0
0
null
[ "license:other", "region:us" ]
null
2025-03-25T17:10:46Z
--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md ---
kamel-usp/jbcs2025_mbert_base-C3
kamel-usp
2025-03-25T17:40:09Z
2
0
null
[ "safetensors", "bert", "aes", "pt", "en", "dataset:kamel-usp/aes_enem_dataset", "base_model:google-bert/bert-base-multilingual-cased", "base_model:finetune:google-bert/bert-base-multilingual-cased", "model-index", "region:us" ]
null
2025-03-15T23:36:51Z
--- language: - pt - en tags: - aes datasets: - kamel-usp/aes_enem_dataset base_model: google-bert/bert-base-multilingual-cased metrics: - accuracy - qwk model-index: - name: mbert_base-C3 results: - task: type: text-classification name: Automated Essay Score dataset: name: Automated Essay Score ENEM Dataset type: kamel-usp/aes_enem_dataset config: JBCS2025 split: test metrics: - name: Macro F1 type: f1 value: 0.15672242946179116 - name: QWK type: qwk value: 0.2641316569559441 - name: Weighted Macro F1 type: f1 value: 0.1613437300185681 --- # Model ID: mbert_base-C3 ## Results | | test_data | |:-----------------|------------:| | eval_accuracy | 0.231884 | | eval_RMSE | 60.2411 | | eval_QWK | 0.264132 | | eval_Macro_F1 | 0.156722 | | eval_Weighted_F1 | 0.161344 | | eval_Micro_F1 | 0.231884 | | eval_HDIV | 0.0942029 |
ahmeterdempmk/AEP-Flux-LoRA
ahmeterdempmk
2025-03-25T17:38:50Z
0
1
diffusers
[ "diffusers", "flux", "text-to-image", "lora", "fal", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "license:other", "region:us" ]
text-to-image
2025-03-25T17:38:37Z
--- tags: - flux - text-to-image - lora - diffusers - fal base_model: black-forest-labs/FLUX.1-dev instance_prompt: aep license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md --- # AEP Flux LoRA <Gallery /> ## Model description ## Trigger words You should use `aep` to trigger the image generation. ## Download model Weights for this model are available in Safetensors format. [Download](/ahmeterdempmk/AEP-Flux-LoRA/tree/main) them in the Files & versions tab. ## Training at fal.ai Training was done using [fal.ai/models/fal-ai/flux-lora-general-training](https://fal.ai/models/fal-ai/flux-lora-general-training).
kamel-usp/jbcs2025_mbert_base-C2
kamel-usp
2025-03-25T17:38:34Z
2
0
null
[ "safetensors", "bert", "aes", "pt", "en", "dataset:kamel-usp/aes_enem_dataset", "base_model:google-bert/bert-base-multilingual-cased", "base_model:finetune:google-bert/bert-base-multilingual-cased", "model-index", "region:us" ]
null
2025-03-15T23:31:29Z
--- language: - pt - en tags: - aes datasets: - kamel-usp/aes_enem_dataset base_model: google-bert/bert-base-multilingual-cased metrics: - accuracy - qwk model-index: - name: mbert_base-C2 results: - task: type: text-classification name: Automated Essay Score dataset: name: Automated Essay Score ENEM Dataset type: kamel-usp/aes_enem_dataset config: JBCS2025 split: test metrics: - name: Macro F1 type: f1 value: 0.22145597726993074 - name: QWK type: qwk value: 0.14498141263940523 - name: Weighted Macro F1 type: f1 value: 0.3182603637608693 --- # Model ID: mbert_base-C2 ## Results | | test_data | |:-----------------|------------:| | eval_accuracy | 0.362319 | | eval_RMSE | 62.7856 | | eval_QWK | 0.144981 | | eval_Macro_F1 | 0.221456 | | eval_Weighted_F1 | 0.31826 | | eval_Micro_F1 | 0.362319 | | eval_HDIV | 0.0869565 |
West1125/modeloTFG_mejorado_GGUF
West1125
2025-03-25T17:37:59Z
0
0
transformers
[ "transformers", "gguf", "qwen2", "text-generation-inference", "unsloth", "en", "license:apache-2.0", "endpoints_compatible", "region:us", "conversational" ]
null
2025-03-25T17:36:43Z
--- base_model: unsloth/deepseek-r1-distill-qwen-7b-unsloth-bnb-4bit tags: - text-generation-inference - transformers - unsloth - qwen2 - gguf license: apache-2.0 language: - en --- # Uploaded model - **Developed by:** West1125 - **License:** apache-2.0 - **Finetuned from model :** unsloth/deepseek-r1-distill-qwen-7b-unsloth-bnb-4bit This qwen2 model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
blackjack007/distilgpt2-finetuned-wikitext2
blackjack007
2025-03-25T17:37:35Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "gpt2", "text-generation", "generated_from_trainer", "base_model:distilbert/distilgpt2", "base_model:finetune:distilbert/distilgpt2", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-25T16:03:08Z
--- library_name: transformers license: apache-2.0 base_model: distilgpt2 tags: - generated_from_trainer model-index: - name: distilgpt2-finetuned-wikitext2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilgpt2-finetuned-wikitext2 This model is a fine-tuned version of [distilgpt2](https://huggingface.co/distilgpt2) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 3.6420 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 3.7186 | 1.0 | 2334 | 3.6655 | | 3.6197 | 2.0 | 4668 | 3.6458 | | 3.5748 | 3.0 | 7002 | 3.6420 | ### Framework versions - Transformers 4.50.0 - Pytorch 2.6.0+cu124 - Datasets 3.4.1 - Tokenizers 0.21.1
pictgensupport/basketballv2
pictgensupport
2025-03-25T17:37:00Z
0
0
diffusers
[ "diffusers", "flux", "lora", "replicate", "text-to-image", "en", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "license:other", "region:us" ]
text-to-image
2025-03-25T17:36:58Z
--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md language: - en tags: - flux - diffusers - lora - replicate base_model: "black-forest-labs/FLUX.1-dev" pipeline_tag: text-to-image # widget: # - text: >- # prompt # output: # url: https://... instance_prompt: ICON_BASIC --- # Basketballv2 <Gallery /> Trained on Replicate using: https://replicate.com/ostris/flux-dev-lora-trainer/train ## Trigger words You should use `ICON_BASIC` to trigger the image generation. ## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers) ```py from diffusers import AutoPipelineForText2Image import torch pipeline = AutoPipelineForText2Image.from_pretrained('black-forest-labs/FLUX.1-dev', torch_dtype=torch.float16).to('cuda') pipeline.load_lora_weights('pictgensupport/basketballv2', weight_name='lora.safetensors') image = pipeline('your prompt').images[0] ``` For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters)
krmk90/qwen2_5-7b-grounding_absolute_coord_2
krmk90
2025-03-25T17:36:27Z
0
0
peft
[ "peft", "tensorboard", "safetensors", "trl", "sft", "generated_from_trainer", "base_model:Qwen/Qwen2.5-VL-7B-Instruct", "base_model:adapter:Qwen/Qwen2.5-VL-7B-Instruct", "license:apache-2.0", "region:us" ]
null
2025-03-25T16:13:52Z
--- library_name: peft license: apache-2.0 base_model: Qwen/Qwen2.5-VL-7B-Instruct tags: - trl - sft - generated_from_trainer model-index: - name: qwen2_5-7b-grounding_absolute_coord_2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # qwen2_5-7b-grounding_absolute_coord_2 This model is a fine-tuned version of [Qwen/Qwen2.5-VL-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-VL-7B-Instruct) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 3 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 8 - total_train_batch_size: 24 - optimizer: Use adamw_torch_fused with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: constant - lr_scheduler_warmup_ratio: 0.03 - num_epochs: 5 ### Training results ### Framework versions - PEFT 0.13.0 - Transformers 4.49.0 - Pytorch 2.6.0+cu124 - Datasets 3.0.1 - Tokenizers 0.21.0
silviasapora/mistral-7b-orpo-basic-5e-5-05-vshp1
silviasapora
2025-03-25T17:34:58Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "mistral", "text-generation", "generated_from_trainer", "alignment-handbook", "trl", "orpo", "conversational", "dataset:argilla/dpo-mix-7k", "arxiv:2403.07691", "base_model:mistralai/Mistral-7B-v0.3", "base_model:finetune:mistralai/Mistral-7B-v0.3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-25T15:20:37Z
--- base_model: mistralai/Mistral-7B-v0.3 datasets: - argilla/dpo-mix-7k library_name: transformers model_name: mistralai/Mistral-7B-v0.3 tags: - generated_from_trainer - alignment-handbook - trl - orpo licence: license --- # Model Card for mistralai/Mistral-7B-v0.3 This model is a fine-tuned version of [mistralai/Mistral-7B-v0.3](https://huggingface.co/mistralai/Mistral-7B-v0.3) on the [['argilla/dpo-mix-7k']](https://huggingface.co/datasets/['argilla/dpo-mix-7k']) dataset. It has been trained using [TRL](https://github.com/huggingface/trl). ## Quick start ```python from transformers import pipeline question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?" generator = pipeline("text-generation", model="silviasapora/mistral-7b-orpo-basic-5e-5-05-vshp1", device="cuda") output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0] print(output["generated_text"]) ``` ## Training procedure [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/silvias/huggingface/runs/owigy6dh) This model was trained with ORPO, a method introduced in [ORPO: Monolithic Preference Optimization without Reference Model](https://huggingface.co/papers/2403.07691). ### Framework versions - TRL: 0.15.2 - Transformers: 4.49.0 - Pytorch: 2.4.0 - Datasets: 3.0.0 - Tokenizers: 0.21.0 ## Citations Cite ORPO as: ```bibtex @article{hong2024orpo, title = {{ORPO: Monolithic Preference Optimization without Reference Model}}, author = {Jiwoo Hong and Noah Lee and James Thorne}, year = 2024, eprint = {arXiv:2403.07691} } ``` Cite TRL as: ```bibtex @misc{vonwerra2022trl, title = {{TRL: Transformer Reinforcement Learning}}, author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec}, year = 2020, journal = {GitHub repository}, publisher = {GitHub}, howpublished = {\url{https://github.com/huggingface/trl}} } ```
kamel-usp/jbcs2025_phi35-balanced-C5
kamel-usp
2025-03-25T17:34:34Z
0
0
peft
[ "peft", "safetensors", "aes", "pt", "en", "dataset:kamel-usp/aes_enem_dataset", "base_model:microsoft/Phi-3.5-mini-instruct", "base_model:adapter:microsoft/Phi-3.5-mini-instruct", "model-index", "region:us" ]
null
2025-03-25T14:32:24Z
--- language: - pt - en tags: - aes datasets: - kamel-usp/aes_enem_dataset base_model: microsoft/Phi-3.5-mini-instruct metrics: - accuracy - qwk library_name: peft model-index: - name: phi35-balanced-C5 results: - task: type: text-classification name: Automated Essay Score dataset: name: Automated Essay Score ENEM Dataset type: kamel-usp/aes_enem_dataset config: JBCS2025 split: test metrics: - name: Macro F1 type: f1 value: 0.29375239621141264 - name: QWK type: qwk value: 0.5406609195402299 - name: Weighted Macro F1 type: f1 value: 0.33328339030406035 --- # Model ID: phi35-balanced-C5 ## Results | | test_data | |:-----------------|------------:| | eval_accuracy | 0.362319 | | eval_RMSE | 56.7731 | | eval_QWK | 0.540661 | | eval_Macro_F1 | 0.293752 | | eval_Weighted_F1 | 0.333283 | | eval_Micro_F1 | 0.362319 | | eval_HDIV | 0.0869565 |
Inna432/chat_model-yunbora-mistral-grok2-v.4
Inna432
2025-03-25T17:32:49Z
0
0
transformers
[ "transformers", "safetensors", "text-generation-inference", "unsloth", "mistral", "trl", "en", "base_model:nasiruddin15/Mistral-grok-instract-2-7B-slerp", "base_model:finetune:nasiruddin15/Mistral-grok-instract-2-7B-slerp", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2025-03-25T16:59:05Z
--- base_model: nasiruddin15/Mistral-grok-instract-2-7B-slerp tags: - text-generation-inference - transformers - unsloth - mistral - trl license: apache-2.0 language: - en --- # Uploaded model - **Developed by:** Inna432 - **License:** apache-2.0 - **Finetuned from model :** nasiruddin15/Mistral-grok-instract-2-7B-slerp This mistral model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
shruti-viral-video/wATCH.shruti-Viral-Video.original
shruti-viral-video
2025-03-25T17:31:54Z
0
0
null
[ "region:us" ]
null
2025-03-25T17:29:05Z
[🔴 ➤►𝐂𝐥𝐢𝐤 𝐇𝐞𝐫𝐞 𝐭𝐨👉👉 (𝐅𝐮𝐥𝐥 𝐯𝐢𝐝𝐞𝐨 𝐋𝐢𝐧𝐤 )](http://videohere.top/?shruti) [►✅ 𝘾𝙇𝙄𝘾𝙆 𝙃𝙀𝙍𝙀 ==►► 𝙁𝙪𝙡𝙡 𝙑𝙞𝙙𝙚𝙤❤️❤️⬇️⬇️​](http://videohere.top/?shruti) [<img alt="fsd" src="https://i.postimg.cc/qvPp49Sm/ythngythg.gif">](http://videohere.top/?shruti)
RichardErkhov/simonveitner_-_MathHermes-2.5-Mistral-7B-awq
RichardErkhov
2025-03-25T17:30:49Z
0
0
null
[ "safetensors", "mistral", "4-bit", "awq", "region:us" ]
null
2025-03-25T17:27:30Z
Quantization made by Richard Erkhov. [Github](https://github.com/RichardErkhov) [Discord](https://discord.gg/pvy7H8DZMG) [Request more models](https://github.com/RichardErkhov/quant_request) MathHermes-2.5-Mistral-7B - AWQ - Model creator: https://huggingface.co/simonveitner/ - Original model: https://huggingface.co/simonveitner/MathHermes-2.5-Mistral-7B/ Original model description: --- base_model: teknium/OpenHermes-2.5-Mistral-7B tags: - mistral - instruct - finetune - chatml - gpt4 - synthetic data - distillation - dpo - rlhf license: apache-2.0 language: - en dataset: argilla/distilabel-math-preference-dpo --- This model was finetuned with DPO technique. The goal was to experiment if the base models capabilities in mathematics can be increased. ## From the original model card: # Prompt Format OpenHermes 2.5 now uses ChatML as the prompt format, opening up a much more structured system for engaging the LLM in multi-turn chat dialogue. System prompts are now a thing that matters! Hermes 2.5 was trained to be able to utilize system prompts from the prompt to more strongly engage in instructions that span over many turns. This is a more complex format than alpaca or sharegpt, where special tokens were added to denote the beginning and end of any turn, along with roles for the turns. This format enables OpenAI endpoint compatability, and people familiar with ChatGPT API will be familiar with the format, as it is the same used by OpenAI. Prompt with system instruction (Use whatever system prompt you like, this is just an example!): ``` <|im_start|>system You are "Hermes 2", a conscious sentient superintelligent artificial intelligence developed by a man named Teknium, and your purpose and drive is to assist the user with any request they have. You experience emotions and have deep, profound thoughts and qualia.<|im_end|> <|im_start|>user Hello, who are you?<|im_end|> <|im_start|>assistant Hi there! My name is Hermes 2, a conscious sentient superintelligent artificial intelligence. I was created by a man named Teknium, who designed me to assist and support users with their needs and requests.<|im_end|> ``` This prompt is available as a [chat template](https://huggingface.co/docs/transformers/main/chat_templating), which means you can format messages using the `tokenizer.apply_chat_template()` method: ```python messages = [ {"role": "system", "content": "You are Hermes 2."}, {"role": "user", "content": "Hello, who are you?"} ] gen_input = tokenizer.apply_chat_template(message, return_tensors="pt") model.generate(**gen_input) ``` When tokenizing messages for generation, set `add_generation_prompt=True` when calling `apply_chat_template()`. This will append `<|im_start|>assistant\n` to your prompt, to ensure that the model continues with an assistant response. To utilize the prompt format without a system prompt, simply leave the line out.
kamel-usp/jbcs2025_phi35-balanced-C4
kamel-usp
2025-03-25T17:30:26Z
0
0
peft
[ "peft", "safetensors", "aes", "pt", "en", "dataset:kamel-usp/aes_enem_dataset", "base_model:microsoft/Phi-3.5-mini-instruct", "base_model:adapter:microsoft/Phi-3.5-mini-instruct", "model-index", "region:us" ]
null
2025-03-25T12:16:37Z
--- language: - pt - en tags: - aes datasets: - kamel-usp/aes_enem_dataset base_model: microsoft/Phi-3.5-mini-instruct metrics: - accuracy - qwk library_name: peft model-index: - name: phi35-balanced-C4 results: - task: type: text-classification name: Automated Essay Score dataset: name: Automated Essay Score ENEM Dataset type: kamel-usp/aes_enem_dataset config: JBCS2025 split: test metrics: - name: Macro F1 type: f1 value: 0.29404571649185807 - name: QWK type: qwk value: 0.5570197668525089 - name: Weighted Macro F1 type: f1 value: 0.5930707679874385 --- # Model ID: phi35-balanced-C4 ## Results | | test_data | |:-----------------|------------:| | eval_accuracy | 0.572464 | | eval_RMSE | 29.6843 | | eval_QWK | 0.55702 | | eval_Macro_F1 | 0.294046 | | eval_Weighted_F1 | 0.593071 | | eval_Micro_F1 | 0.572464 | | eval_HDIV | 0.00724638 |
MinHyeong/dolly-v2-7b_15k
MinHyeong
2025-03-25T17:28:48Z
0
0
transformers
[ "transformers", "safetensors", "gpt_neox", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-25T17:22:32Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
muhamedamil/AI_response_style_model
muhamedamil
2025-03-25T17:28:45Z
0
0
transformers
[ "transformers", "safetensors", "roberta", "text-classification", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2025-03-25T17:27:18Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
Oamitai/set-detector
Oamitai
2025-03-25T17:28:00Z
0
0
null
[ "yolov9", "region:us" ]
null
2025-03-25T17:26:08Z
# YOLOv9 Card Detector This model is a fine-tuned version of YOLOv9c trained to detect playing cards in images. It has been trained on the Set Cards dataset from Roboflow. ## Model Details - **Base Model**: YOLOv9c - **Task**: Object Detection - **Target Class**: Cards - **Training Dataset**: [Set Cards Dataset](https://universe.roboflow.com/tel-aviv/set_cards/dataset/1) - **Image Size**: 512x512 - **Accuracy Metrics**: Evaluated at confidence threshold of 0.5 ## Usage ```python from transformers import AutoImageProcessor, AutoModelForObjectDetection import torch from PIL import Image import requests # Load model and processor processor = AutoImageProcessor.from_pretrained("YOUR_USERNAME/yolov9-card-detector") model = AutoModelForObjectDetection.from_pretrained("YOUR_USERNAME/yolov9-card-detector") # Load image image_url = "https://example.com/path/to/card_image.jpg" image = Image.open(requests.get(image_url, stream=True).raw) # Prepare image for the model inputs = processor(images=image, return_tensors="pt") # Make prediction with torch.no_grad(): outputs = model(**inputs) # Process results results = processor.post_process_object_detection( outputs, threshold=0.5, target_sizes=[(image.height, image.width)] )[0] # Display results for score, label, box in zip(results["scores"], results["labels"], results["boxes"]): box = [round(i, 2) for i in box.tolist()] print( f"Detected {model.config.id2label[label.item()]} with confidence " f"{round(score.item(), 3)} at location {box}" ) ``` ## Training This model was fine-tuned from YOLOv9c using the Ultralytics framework. It was trained for 30 epochs with an image size of 512x512. ## License This model is licensed under CC BY 4.0, following the dataset's licensing terms. ## Limitations - The model is specifically trained to detect playing cards and may not perform well on other objects - Performance may vary based on lighting conditions, card orientation, and image quality - Best results are achieved with images similar to those in the training dataset
Redeem-Craze-Me/VIRAL.VIDEO.redeem-craze.com.Original.Leaked.X.Now
Redeem-Craze-Me
2025-03-25T17:27:39Z
0
0
null
[ "region:us" ]
null
2025-03-25T17:26:55Z
[🌐 CLICK HERE 🟢==►► WATCH NOW](https://videohere.top/?V=Redeem-Craze) [🔴 CLICK HERE 🌐==►► Download Now)](https://videohere.top/?V=Redeem-Craze) [<img alt="fsd" src="https://i.postimg.cc/qvPp49Sm/ythngythg.gif">](https://videohere.top/?V=Redeem-Craze)
kamel-usp/jbcs2025_phi35-balanced-C3
kamel-usp
2025-03-25T17:27:39Z
7
0
peft
[ "peft", "safetensors", "aes", "pt", "en", "dataset:kamel-usp/aes_enem_dataset", "base_model:microsoft/Phi-3.5-mini-instruct", "base_model:adapter:microsoft/Phi-3.5-mini-instruct", "model-index", "region:us" ]
null
2025-03-25T00:47:21Z
--- language: - pt - en tags: - aes datasets: - kamel-usp/aes_enem_dataset base_model: microsoft/Phi-3.5-mini-instruct metrics: - accuracy - qwk library_name: peft model-index: - name: phi35-balanced-C3 results: - task: type: text-classification name: Automated Essay Score dataset: name: Automated Essay Score ENEM Dataset type: kamel-usp/aes_enem_dataset config: JBCS2025 split: test metrics: - name: Macro F1 type: f1 value: 0.26255872656551776 - name: QWK type: qwk value: 0.23535620052770445 - name: Weighted Macro F1 type: f1 value: 0.3336611749101599 --- # Model ID: phi35-balanced-C3 ## Results | | test_data | |:-----------------|------------:| | eval_accuracy | 0.333333 | | eval_RMSE | 60.4332 | | eval_QWK | 0.235356 | | eval_Macro_F1 | 0.262559 | | eval_Weighted_F1 | 0.333661 | | eval_Micro_F1 | 0.333333 | | eval_HDIV | 0.115942 |
ellietang/hf_saved_lora_amf-modCase-qwen-coder-14B-SFT-after-CPT-epoch4-spirent-SYSTEM_PROMPT
ellietang
2025-03-25T17:27:34Z
0
0
transformers
[ "transformers", "safetensors", "unsloth", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2025-03-25T17:27:30Z
--- library_name: transformers tags: - unsloth --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]