modelId
stringlengths
5
139
author
stringlengths
2
42
last_modified
timestamp[us, tz=UTC]date
2020-02-15 11:33:14
2025-06-24 12:28:46
downloads
int64
0
223M
likes
int64
0
11.7k
library_name
stringclasses
493 values
tags
sequencelengths
1
4.05k
pipeline_tag
stringclasses
54 values
createdAt
timestamp[us, tz=UTC]date
2022-03-02 23:29:04
2025-06-24 12:27:57
card
stringlengths
11
1.01M
PyWebSol/PySols-OCR-DETR
PyWebSol
2025-03-07T12:47:27Z
0
0
transformers
[ "transformers", "safetensors", "detr", "object-detection", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
object-detection
2025-03-07T10:13:06Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
Able2/umt5-xxl-encode-only
Able2
2025-03-07T12:47:22Z
0
0
transformers
[ "transformers", "safetensors", "gguf", "umt5", "text2text-generation", "base_model:google/umt5-xxl", "base_model:quantized:google/umt5-xxl", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text2text-generation
2025-03-07T12:11:39Z
--- license: apache-2.0 base_model: - google/umt5-xxl library_name: transformers --- An encoder only version of google's umt5-xxl model, mainly used as text encoder for image or video generation models. Original Repo: https://huggingface.co/google/umt5-xxl Quantized version: https://huggingface.co/Able2/umt5-xxl-encode-only-gguf
marcuscedricridia/Yell-Qwen2.5-7B-1M-della1
marcuscedricridia
2025-03-07T12:46:25Z
0
0
transformers
[ "transformers", "safetensors", "qwen2", "text-generation", "mergekit", "merge", "conversational", "arxiv:2406.11617", "base_model:marcuscedricridia/Yell-Qwen2.5-7B-1M", "base_model:merge:marcuscedricridia/Yell-Qwen2.5-7B-1M", "base_model:nvidia/AceInstruct-7B", "base_model:merge:nvidia/AceInstruct-7B", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-07T12:43:07Z
--- base_model: - marcuscedricridia/Yell-Qwen2.5-7B-1M - nvidia/AceInstruct-7B library_name: transformers tags: - mergekit - merge --- # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [DELLA](https://arxiv.org/abs/2406.11617) merge method using [marcuscedricridia/Yell-Qwen2.5-7B-1M](https://huggingface.co/marcuscedricridia/Yell-Qwen2.5-7B-1M) as a base. ### Models Merged The following models were included in the merge: * [nvidia/AceInstruct-7B](https://huggingface.co/nvidia/AceInstruct-7B) ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: nvidia/AceInstruct-7B parameters: density: 1 weight: 1 lambda: 0.9 merge_method: della base_model: marcuscedricridia/Yell-Qwen2.5-7B-1M parameters: density: 1 weight: 1 lambda: 0.9 normalize: true int8_mask: true dtype: bfloat16 tokenizer_source: base name: Yell-Qwen2.5-7B-1M-della1 ```
bakhtawar2304/quran-recitation-model2
bakhtawar2304
2025-03-07T12:45:38Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "whisper", "automatic-speech-recognition", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
automatic-speech-recognition
2025-03-07T07:04:16Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
lesso14/d2e9e20f-38e6-4321-aa18-d2b310d8b25a
lesso14
2025-03-07T12:44:34Z
0
0
peft
[ "peft", "safetensors", "llama", "axolotl", "generated_from_trainer", "custom_code", "base_model:NousResearch/Yarn-Llama-2-13b-128k", "base_model:adapter:NousResearch/Yarn-Llama-2-13b-128k", "region:us" ]
null
2025-03-07T10:00:58Z
--- library_name: peft base_model: NousResearch/Yarn-Llama-2-13b-128k tags: - axolotl - generated_from_trainer model-index: - name: d2e9e20f-38e6-4321-aa18-d2b310d8b25a results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.4.1` ```yaml adapter: lora base_model: NousResearch/Yarn-Llama-2-13b-128k bf16: auto chat_template: llama3 dataset_prepared_path: null datasets: - data_files: - f5e446c398dddd6a_train_data.json ds_type: json format: custom path: /workspace/input_data/f5e446c398dddd6a_train_data.json type: field_input: plan field_instruction: goal field_output: revision format: '{instruction} {input}' no_input_format: '{instruction}' system_format: '{system}' system_prompt: '' debug: null deepspeed: null do_eval: true early_stopping_patience: 3 eval_batch_size: 4 eval_max_new_tokens: 128 eval_steps: 500 evals_per_epoch: null flash_attention: false fp16: false fsdp: null fsdp_config: null gradient_accumulation_steps: 8 gradient_checkpointing: true group_by_length: true hub_model_id: lesso14/d2e9e20f-38e6-4321-aa18-d2b310d8b25a hub_repo: null hub_strategy: checkpoint hub_token: null learning_rate: 0.000214 load_in_4bit: false load_in_8bit: false local_rank: null logging_steps: 50 lora_alpha: 128 lora_dropout: 0.15 lora_fan_in_fan_out: null lora_model_dir: null lora_r: 64 lora_target_linear: true lr_scheduler: cosine max_grad_norm: 1.0 max_steps: 500 micro_batch_size: 4 mlflow_experiment_name: /tmp/f5e446c398dddd6a_train_data.json model_type: AutoModelForCausalLM num_epochs: 10 optimizer: adamw_torch_fused output_dir: miner_id_24 pad_to_sequence_len: true resume_from_checkpoint: null s2_attention: null sample_packing: false save_steps: 500 saves_per_epoch: null seed: 140 sequence_len: 1024 strict: false tf32: true tokenizer_type: AutoTokenizer train_on_inputs: false trust_remote_code: true val_set_size: 0.05 wandb_entity: null wandb_mode: online wandb_name: b24bdad9-e2c6-4a01-8b85-11c2aeda3b8a wandb_project: 14a wandb_run: your_name wandb_runid: b24bdad9-e2c6-4a01-8b85-11c2aeda3b8a warmup_steps: 100 weight_decay: 0.0 xformers_attention: null ``` </details><br> # d2e9e20f-38e6-4321-aa18-d2b310d8b25a This model is a fine-tuned version of [NousResearch/Yarn-Llama-2-13b-128k](https://huggingface.co/NousResearch/Yarn-Llama-2-13b-128k) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.9641 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.000214 - train_batch_size: 4 - eval_batch_size: 4 - seed: 140 - gradient_accumulation_steps: 8 - total_train_batch_size: 32 - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 100 - training_steps: 500 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:------:|:----:|:---------------:| | No log | 0.0006 | 1 | 1.6201 | | 7.6292 | 0.2802 | 500 | 0.9641 | ### Framework versions - PEFT 0.13.2 - Transformers 4.46.0 - Pytorch 2.5.0+cu124 - Datasets 3.0.1 - Tokenizers 0.20.1
XuehangCang/autotrain-u9u6w-ehmyh
XuehangCang
2025-03-07T12:43:59Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "autotrain", "text-generation-inference", "text-generation", "peft", "conversational", "dataset:XuehangCang/jianke", "base_model:Qwen/QwQ-32B", "base_model:finetune:Qwen/QwQ-32B", "license:other", "endpoints_compatible", "region:us" ]
text-generation
2025-03-07T07:13:58Z
--- tags: - autotrain - text-generation-inference - text-generation - peft library_name: transformers base_model: Qwen/QwQ-32B widget: - messages: - role: user content: What is your favorite condiment? license: other datasets: - XuehangCang/jianke --- # Model Trained Using AutoTrain This model was trained using AutoTrain. For more information, please visit [AutoTrain](https://hf.co/docs/autotrain). # Usage ```python from transformers import AutoModelForCausalLM, AutoTokenizer model_path = "PATH_TO_THIS_REPO" tokenizer = AutoTokenizer.from_pretrained(model_path) model = AutoModelForCausalLM.from_pretrained( model_path, device_map="auto", torch_dtype='auto' ).eval() # Prompt content: "hi" messages = [ {"role": "user", "content": "hi"} ] input_ids = tokenizer.apply_chat_template(conversation=messages, tokenize=True, add_generation_prompt=True, return_tensors='pt') output_ids = model.generate(input_ids.to('cuda')) response = tokenizer.decode(output_ids[0][input_ids.shape[1]:], skip_special_tokens=True) # Model response: "Hello! How can I assist you today?" print(response) ```
amrithhun/Qwen_Atlas
amrithhun
2025-03-07T12:43:55Z
0
0
null
[ "license:apache-2.0", "region:us" ]
null
2025-03-07T12:43:55Z
--- license: apache-2.0 ---
lesso11/baf27499-cd61-4529-ad1c-290046209573
lesso11
2025-03-07T12:42:13Z
0
0
peft
[ "peft", "safetensors", "llama", "axolotl", "generated_from_trainer", "custom_code", "base_model:NousResearch/CodeLlama-13b-hf-flash", "base_model:adapter:NousResearch/CodeLlama-13b-hf-flash", "region:us" ]
null
2025-03-07T09:30:54Z
--- library_name: peft base_model: NousResearch/CodeLlama-13b-hf-flash tags: - axolotl - generated_from_trainer model-index: - name: baf27499-cd61-4529-ad1c-290046209573 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.4.1` ```yaml adapter: lora base_model: NousResearch/CodeLlama-13b-hf-flash bf16: auto chat_template: llama3 dataset_prepared_path: null datasets: - data_files: - a4ca02b335779d76_train_data.json ds_type: json format: custom path: /workspace/input_data/a4ca02b335779d76_train_data.json type: field_instruction: premise field_output: hypothesis format: '{instruction}' no_input_format: '{instruction}' system_format: '{system}' system_prompt: '' debug: null deepspeed: null do_eval: true early_stopping_patience: 3 eval_batch_size: 4 eval_max_new_tokens: 128 eval_steps: 500 evals_per_epoch: null flash_attention: false fp16: false fsdp: null fsdp_config: null gradient_accumulation_steps: 8 gradient_checkpointing: true group_by_length: true hub_model_id: lesso11/baf27499-cd61-4529-ad1c-290046209573 hub_repo: null hub_strategy: checkpoint hub_token: null learning_rate: 0.000211 load_in_4bit: false load_in_8bit: false local_rank: null logging_steps: 50 lora_alpha: 128 lora_dropout: 0.15 lora_fan_in_fan_out: null lora_model_dir: null lora_r: 64 lora_target_linear: true lr_scheduler: cosine max_grad_norm: 1.0 max_steps: 500 micro_batch_size: 4 mlflow_experiment_name: /tmp/a4ca02b335779d76_train_data.json model_type: AutoModelForCausalLM num_epochs: 10 optimizer: adamw_torch_fused output_dir: miner_id_24 pad_to_sequence_len: true resume_from_checkpoint: null s2_attention: null sample_packing: false save_steps: 500 saves_per_epoch: null seed: 110 sequence_len: 1024 special_tokens: pad_token: </s> strict: false tf32: true tokenizer_type: AutoTokenizer train_on_inputs: false trust_remote_code: true val_set_size: 0.05 wandb_entity: null wandb_mode: online wandb_name: b6ec0ba2-3f0e-46c4-8a4c-3ad9a8d7355f wandb_project: 11a wandb_run: your_name wandb_runid: b6ec0ba2-3f0e-46c4-8a4c-3ad9a8d7355f warmup_steps: 100 weight_decay: 0.0 xformers_attention: null ``` </details><br> # baf27499-cd61-4529-ad1c-290046209573 This model is a fine-tuned version of [NousResearch/CodeLlama-13b-hf-flash](https://huggingface.co/NousResearch/CodeLlama-13b-hf-flash) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.5641 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.000211 - train_batch_size: 4 - eval_batch_size: 4 - seed: 110 - gradient_accumulation_steps: 8 - total_train_batch_size: 32 - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 100 - training_steps: 500 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:------:|:----:|:---------------:| | No log | 0.0002 | 1 | 2.7520 | | 12.3741 | 0.1002 | 500 | 1.5641 | ### Framework versions - PEFT 0.13.2 - Transformers 4.46.0 - Pytorch 2.5.0+cu124 - Datasets 3.0.1 - Tokenizers 0.20.1
nallarahul/NewsGaurd
nallarahul
2025-03-07T12:37:03Z
0
0
transformers
[ "transformers", "safetensors", "bert", "text-classification", "fake-news-detection", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2025-02-25T18:14:38Z
--- language: en license: apache-2.0 tags: - fake-news-detection - bert - text-classification - transformers --- # NewsGuard AI - Fake News Detection Model This model is a fine-tuned **BERT-base-uncased** model for detecting fake news. It is trained using the **FakeNewsNet** dataset. ## Model Details - **Base Model:** BERT-base-uncased - **Task:** Text Classification (Fake vs. Real News) - **Dataset:** FakeNewsNet (GossipCop & PolitiFact) - **Training Framework:** Hugging Face Transformers - **Metrics:** Accuracy, Precision, Recall ## How to Use ```python from transformers import AutoModelForSequenceClassification, AutoTokenizer import torch model_path = "your-huggingface-username/newsguard-ai-fake-news" tokenizer = AutoTokenizer.from_pretrained(model_path) model = AutoModelForSequenceClassification.from_pretrained(model_path) text = "Some news article text here..." inputs = tokenizer(text, return_tensors="pt", truncation=True, padding=True, max_length=512) with torch.no_grad(): outputs = model(**inputs) probs = torch.nn.functional.softmax(outputs.logits, dim=-1) prediction = "Fake" if torch.argmax(probs) == 0 else "Real" print(f"Prediction: {prediction}, Confidence: {probs.tolist()[0]}")
idopinto/co-specter2-biomed
idopinto
2025-03-07T12:32:44Z
17
0
transformers
[ "transformers", "pytorch", "bert", "feature-extraction", "arxiv:2111.08366", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2025-02-17T09:31:04Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> SPECTER-CoCite model included in a paper for modeling fine-grained similarity between documents in the biomedical domain. Title: "Multi-Vector Models with Textual Guidance for Fine-Grained Scientific Document Similarity" Authors: Sheshera Mysore, Arman Cohan, Tom Hope Paper: https://arxiv.org/abs/2111.08366 Github: https://github.com/allenai/aspire Note: In the context of the paper, this model is referred to as Specter-CoCite_Spec and represents a baseline bi-encoder for scientific document similarity. This model is similar in architecture to the allenai/specter model but is trained on co-citation data instead of citation data. Refer to https://huggingface.co/allenai/aspire-biencoder-biomed-spec for more details and usage. Base Model: https://huggingface.co/allenai/specter2_base ## Evaluation Results Differences might be in co-citations training data which constructed from scratch from different release of S2ORC (originaly, 2019-09-28, which I didn't have access to.) | Model | TRECCOVID-MAP | TRECCOVID-NDCG%20 | RELISH-MAP | RELISH-NDCG%20 | |--------------------------------|----------|----------|----------|----------| | specter | 28.24 | 59.28 |60.62 | 77.20 | | aspire-biencoder-biomed-spec | 26.07 | 54.89 | 61.47 | 78.34 | | aspire-biencoder-biomed-spec-full | 28.87 | 60.47 | 61.69 | 78.22 | | aspire-biencoder-biomed-spec-recon | 27.73 | 59.18 | 60.36 | 77.09 | | **aspire-biencoder-biomed-spec2** | 29.38 | 61.45 | 60.26 | 77.25 |
Enzo121/enzo
Enzo121
2025-03-07T12:32:08Z
0
0
diffusers
[ "diffusers", "flux", "lora", "replicate", "text-to-image", "en", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "license:other", "region:us" ]
text-to-image
2025-03-07T12:01:56Z
--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md language: - en tags: - flux - diffusers - lora - replicate base_model: "black-forest-labs/FLUX.1-dev" pipeline_tag: text-to-image # widget: # - text: >- # prompt # output: # url: https://... instance_prompt: enzo --- # Enzo <Gallery /> Trained on Replicate using: https://replicate.com/ostris/flux-dev-lora-trainer/train ## Trigger words You should use `enzo ` to trigger the image generation. ## Use it with the [🧨 diffusers library](https://github.com/huggingface/diffusers) ```py from diffusers import AutoPipelineForText2Image import torch pipeline = AutoPipelineForText2Image.from_pretrained('black-forest-labs/FLUX.1-dev', torch_dtype=torch.float16).to('cuda') pipeline.load_lora_weights('Enzo121/enzo', weight_name='lora.safetensors') image = pipeline('your prompt').images[0] ``` For more details, including weighting, merging and fusing LoRAs, check the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters)
xkaska02/lilt-robeczech-base
xkaska02
2025-03-07T12:32:00Z
0
0
transformers
[ "transformers", "pytorch", "roberta", "fill-mask", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
fill-mask
2025-03-06T21:50:10Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
just-ne-just/working
just-ne-just
2025-03-07T12:30:43Z
0
0
transformers
[ "transformers", "safetensors", "llama", "text-classification", "generated_from_trainer", "trl", "reward-trainer", "dataset:HumanLLMs/Human-Like-DPO-Dataset", "base_model:HuggingFaceTB/SmolLM-135M-Instruct", "base_model:finetune:HuggingFaceTB/SmolLM-135M-Instruct", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-classification
2025-03-07T12:30:09Z
--- base_model: HuggingFaceTB/SmolLM-135M-Instruct datasets: HumanLLMs/Human-Like-DPO-Dataset library_name: transformers model_name: working tags: - generated_from_trainer - trl - reward-trainer licence: license --- # Model Card for working This model is a fine-tuned version of [HuggingFaceTB/SmolLM-135M-Instruct](https://huggingface.co/HuggingFaceTB/SmolLM-135M-Instruct) on the [HumanLLMs/Human-Like-DPO-Dataset](https://huggingface.co/datasets/HumanLLMs/Human-Like-DPO-Dataset) dataset. It has been trained using [TRL](https://github.com/huggingface/trl). ## Quick start ```python from transformers import pipeline question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?" generator = pipeline("text-generation", model="just-ne-just/working", device="cuda") output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0] print(output["generated_text"]) ``` ## Training procedure This model was trained with Reward. ### Framework versions - TRL: 0.15.2 - Transformers: 4.47.0 - Pytorch: 2.5.1+cu121 - Datasets: 3.3.1 - Tokenizers: 0.21.0 ## Citations Cite TRL as: ```bibtex @misc{vonwerra2022trl, title = {{TRL: Transformer Reinforcement Learning}}, author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec}, year = 2020, journal = {GitHub repository}, publisher = {GitHub}, howpublished = {\url{https://github.com/huggingface/trl}} } ```
jgayed/lorafull120-F16-GGUF
jgayed
2025-03-07T12:28:13Z
0
0
peft
[ "peft", "gguf", "llama-factory", "lora", "generated_from_trainer", "llama-cpp", "gguf-my-lora", "base_model:jgayed/lorafull120", "base_model:adapter:jgayed/lorafull120", "license:other", "region:us" ]
null
2025-03-07T12:28:02Z
--- library_name: peft license: other base_model: jgayed/lorafull120 tags: - llama-factory - lora - generated_from_trainer - llama-cpp - gguf-my-lora model-index: - name: train3 results: [] --- # jgayed/lorafull120-F16-GGUF This LoRA adapter was converted to GGUF format from [`jgayed/lorafull120`](https://huggingface.co/jgayed/lorafull120) via the ggml.ai's [GGUF-my-lora](https://huggingface.co/spaces/ggml-org/gguf-my-lora) space. Refer to the [original adapter repository](https://huggingface.co/jgayed/lorafull120) for more details. ## Use with llama.cpp ```bash # with cli llama-cli -m base_model.gguf --lora lorafull120-f16.gguf (...other args) # with server llama-server -m base_model.gguf --lora lorafull120-f16.gguf (...other args) ``` To know more about LoRA usage with llama.cpp server, refer to the [llama.cpp server documentation](https://github.com/ggerganov/llama.cpp/blob/master/examples/server/README.md).
kaweizhenpi/SmolGRPO-135M
kaweizhenpi
2025-03-07T12:20:24Z
0
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "trl", "grpo", "GRPO", "Reasoning-Course", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-07T12:19:03Z
--- library_name: transformers tags: - trl - grpo - GRPO - Reasoning-Course --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
idopinto/ts-aspire-biomed-specter2
idopinto
2025-03-07T12:18:41Z
15
0
transformers
[ "transformers", "safetensors", "bert", "feature-extraction", "arxiv:2111.08366", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
feature-extraction
2025-02-08T07:13:13Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> tsAspire model included in a paper for modeling fine-grained similarity between documents in the biomedical domain fine-tuned from Specter2-base. Title: "Multi-Vector Models with Textual Guidance for Fine-Grained Scientific Document Similarity" Authors: Sheshera Mysore, Arman Cohan, Tom Hope Paper: https://arxiv.org/abs/2111.08366 Github: https://github.com/allenai/aspire Note: In the context of the paper, this model is referred to as tsAspire and represents the papers proposed multi-vector model for fine-grained scientific document similarity. Refer to https://huggingface.co/allenai/aspire-contextualsentence-singlem-biomed for more details and usage. Base Model: https://huggingface.co/allenai/specter2_base ## Evaluation Results Differences might be in co-citations training data which constructed from scratch from different release of S2ORC (originaly, 2019-09-28, which I didn't have access to.) | Model | Specification | TRECCOVID-MAP | TRECCOVID-NDCG%20 | RELISH-MAP | RELISH-NDCG%20 | |--------------------------------|------------------|----------|----------|----------|----------| |aspire-contextualsentence-singlem-biomed | TSASPIRE_spec_orig | 26.68 | 57.21 |61.06 | 77.20 | | ts-aspire-biomed-recon | TSASPIRE_Spec | 29.26 | 60.45 | 62.2 | 78.7 | | **ts-aspire-biomed-specter2** | TSASPIRE_Spec2 | 31.16 | 62.43 | 63.24 | 79.89 |
techdkit/Gwalior-call-girls-service-7290901024
techdkit
2025-03-07T12:18:30Z
0
0
null
[ "region:us" ]
null
2025-03-07T12:18:13Z
<p>Looking for the&nbsp;<strong>best Russian escorts in Gwalior</strong>? Experience ultimate pleasure and companionship with our&nbsp;<strong>high-class call girls</strong>, available in&nbsp;<strong>Aerocity,&nbsp;Gwalior, and luxury hotels across Gwalior</strong>. Whether you seek a passionate night, an elegant date, or a sensual massage, our&nbsp;<strong>premium Russian escorts</strong>&nbsp;ensure an unforgettable experience.&nbsp;100%&nbsp;<strong>discreet &amp; safe services</strong></p> <h3><strong><a href="https://api.whatsapp.com/send/?phone=917290901024&amp;text&amp;type=phone_number&amp;app_absent=0">WhatsApp Me: 7290901024</a></strong></h3> <p><strong>Visit Website &gt; <a href="https://gwalior.noida-escort.com/">https://gwalior.noida-escort.com</a></strong></p> <p>Call Girls girl in Gwalior&nbsp;call girls Call Girl service in Gwalior&nbsp;Gwalior Call Girls Locanto&nbsp;call girls Call Girl service in Gwalior&nbsp;Gurgaonsex Call Girls&nbsp;Escorts Service at Gwalior&nbsp;Gwalior Escort Service&nbsp;Gwalior Escort girl&nbsp;Escort service Gwalior&nbsp;female Escort in Gwalior&nbsp;Gurgaonsex girl&nbsp;Call Girls in Gwalior&nbsp;Call Girls services Gwalior&nbsp;Gwalior female Escorts&nbsp;housewife Escort in Gwalior&nbsp;Gwalior Escort&nbsp;Gwalior Escorts&nbsp;Gwalior Escort service&nbsp;model Escorts in Gwalior&nbsp;Gwalior Call Girls services&nbsp;Call Girls service Gwalior&nbsp;Russian Escort in Gwalior&nbsp;call girls near me&nbsp;Gwalior Escoorts Call Girls&nbsp;royal Call Girls in Gwalior&nbsp;Gwalior independent Escort&nbsp;cheap Gwalior Escort&nbsp;Escort girls Service Gwalior&nbsp;independent Gwalior Call Girls&nbsp;foreign Escort Girls in Gwalior&nbsp;cheap female Escort Girls in Gwalior&nbsp;royal Call Girls Gwalior&nbsp;hotel Call Girls in Gwalior&nbsp;Gwalior Call Girls girls&nbsp;female Call Girl Gwalior&nbsp;sex Call Girls in Gwalior&nbsp;Call Girl service near me&nbsp;call girls&nbsp;Russian Call Girls Gwalior&nbsp;independent Call Girls in Gwalior&nbsp;cheap Call Girls in Gwalior&nbsp;model Call Girls Gwalior&nbsp;Call Girls service in Gwalior&nbsp;Call Girls Gwalior&nbsp;female Call Girls Gwalior&nbsp;Call Girls services in Gwalior&nbsp;GurgaonVIP Call Girls&nbsp;call girl near me&nbsp;Gurgaonsex&nbsp;Gurgaoncall girl number&nbsp;independent Call Girls Gwalior&nbsp;cheap Call Girls Gwalior&nbsp;girls Call Girl Gwalior&nbsp;call girls Call Girls service in Gwalior&nbsp;Call Girls service in Gwalior&nbsp;independent call girls&nbsp;VIP Call Girls in Gwalior&nbsp;Gurgaonmodel Call Girls&nbsp;female Call Girls of Gwalior&nbsp;Gwalior Call Girl agency&nbsp;Gwalior Call Girl service&nbsp;call girls Call Girl service in Gwalior&nbsp;GurgaonRussian Call Girls&nbsp;Gurgaoni sex video&nbsp;call girl number&nbsp;call girls Gwalior&nbsp;Gwalior Call Girls agency&nbsp;call girls Call Girl service in Gwalior&nbsp;Gurgaonsex&nbsp;call girl contact number&nbsp;sexy Call Girls call girls in Gwalior&nbsp;Gwalior Call Girls&nbsp;कॉल गर्ल लिस्ट Gwalior&nbsp;call girls Call Girl service in Gwalior&nbsp;call girl in Gwalior&nbsp;best Call Girls in Gwalior&nbsp;Call Girls Gwalior</p>
pharci/anime-akira-flux
pharci
2025-03-07T12:16:55Z
29
0
diffusers
[ "diffusers", "flux", "lora", "anime", "text-to-image", "en", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "license:other", "region:us" ]
text-to-image
2025-03-01T17:10:33Z
--- license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md language: - en tags: - flux - diffusers - lora - anime base_model: black-forest-labs/FLUX.1-dev pipeline_tag: text-to-image instance_prompt: AKI widget: - text: >- aki Fantasy, A mystical, cosmic scene bathed in shades of deep violet and purple. The name 'Akira' is formed from a swirling vortex of violet energy in the sky, with glowing, translucent letters that seem to pulse with life. These letters are surrounded by ethereal, shimmering particles of light that float in the air, casting a radiant, soft glow across the surroundings. Below, a tranquil, reflective lake mirrors the cosmic display above, with floating islands draped in soft violet mist. The sky is filled with swirling clouds in hues of purple and indigo, and distant stars twinkle faintly. The lighting is dramatic, with the swirling vortex casting vivid purple and blue light onto the landscape, creating sharp contrasts and mysterious shadows. output: url: images/out-0 (4).png - text: >- AKI Action,close up of a dark, powerful figure in the middle of a fierce battle. His sharp, piercing eyes glow with intense red fury, and his face is set in an expression of unyielding determination. He wields a massive sword that hums with dark energy. His black hair is spiked, and his muscular body shows the toll of the fight. The camera angle is slightly below him, emphasizing his overwhelming strength. The lighting is dramatic casting deep shadows on his face and armor, while his dark aura pulses brightly against the backdrop of a shattered city filled with destruction and chaos output: url: images/out-0 (2).png - text: >- AKI Action, A dynamic action scene in a neon-lit city, a young hero with spiky blue hair and a red jacket, wielding a glowing katana. He’s surrounded by futuristic skyscrapers, with a glowing moon in the background. His eyes are intense, filled with determination. A robotic enemy with glowing orange eyes stands before him, ready to fight. The scene is framed with a low-angle shot, high contrast between the vibrant neon lights and dark shadows, with a strong focus on the characters' expressions and movement. output: url: images/example_ga9d3glfi.png - text: >- AKI Romance, a peaceful moment with a young girl reflecting by the lake, surrounded by nature and tranquility. She has long, flowing pink hair and is wearing a pastel-colored dress, sitting gracefully on a wooden dock. Her expression is soft and peaceful, with a gentle smile on her face. The scene is framed with a medium shot focusing on her figure and the surrounding peaceful environment. The setting is a calm lake with cherry blossom trees in full bloom, soft ripples on the water, and a sunset sky with warm colors. The lighting is soft golden from the setting sun, creating gentle reflections on the water and casting a warm glow on the girl, with subtle contrasts to give a peaceful atmosphere. output: url: images/example_ofi4bzc73.png - text: >- AKI character, a young boy with short, spiky black hair and bright green eyes, looking directly at the viewer with a confident yet calm expression. He’s wearing a dark hoodie and simple jeans, with his arms crossed in front of him. The scene is framed as a close-up shot, focusing on his face and upper body, with the background blurred to keep the attention on him. The setting is a soft gradient of blue and purple, giving a peaceful yet mysterious vibe. The lighting is soft and moody, with gentle highlights on his face, creating depth and emphasizing his calm but confident expression output: url: images/example_fyuascl87.png - text: >- AKI fantasy, A magical forest illuminated by ethereal, glowing plants. A young mage with long silver hair and a dark purple cloak stands in front of an ancient stone archway. She holds a staff with a crystal orb that pulses with light, casting a soft glow. In the background, a massive dragon with shimmering scales flies between the trees, its wings spread wide. The scene is a wide shot, with a focus on the magical energy and mystical creatures, vibrant colors and soft lighting with a dreamlike atmosphere output: url: images/example_g9skimkab.png --- # Akira <Gallery /> ## Instance Prompt Use `AKI` to trigger the image generation. ## How to Use with the [🧨 Diffusers Library](https://github.com/huggingface/diffusers) ```python from diffusers import AutoPipelineForText2Image import torch pipeline = AutoPipelineForText2Image.from_pretrained('black-forest-labs/FLUX.1-dev', torch_dtype=torch.float16).to('cuda') pipeline.load_lora_weights('pharci/akira', weight_name='lora.safetensors') image = pipeline('your prompt').images[0] ``` ## Prompt Tutorial To achieve the desired results with the model, formulate your prompts using the following structure: - **Genre**: Action, Romance, Drama, etc. - **Description**: What’s happening, character dynamics, etc. - **Characters**: Appearance, emotions, clothing, poses, interaction style, etc. - **Framing**: Close-up, medium shot, angle, focus details, etc. - **Setting**: Dark, bright, dramatic, city lights, sunset, etc. - **Lighting**: Simplified backgrounds, vibrant colors, sharp contours, etc. ### Example Prompt "Fantasy, A dramatic battle scene between a brave knight and a fierce dragon, The knight in shining armor, showing determination, while the dragon breathes fire, low-angle shot, set in a dark forest illuminated by firelight, vibrant colors with sharp contours." ## Additional Information This model is based on the FLUX architecture and has been adapted using LoRA weights trained on a custom dataset consisting of 500 images. The dataset focuses on specific visual styles and characteristics. **Note:** This is an early test version of the model and not a final, polished release. It is still a work in progress, and future versions may include improvements and refinements. For more details, including weighting, merging, and fusing LoRAs, refer to the [documentation on loading LoRAs in diffusers](https://huggingface.co/docs/diffusers/main/en/using-diffusers/loading_adapters). ---
EvgenyBondarenko/BiEncoderRanker
EvgenyBondarenko
2025-03-07T12:15:58Z
0
0
sentence-transformers
[ "sentence-transformers", "safetensors", "xlm-roberta", "sentence-similarity", "feature-extraction", "generated_from_trainer", "dataset_size:79515", "loss:ContrastiveLoss", "arxiv:1908.10084", "base_model:deepvk/USER-bge-m3", "base_model:finetune:deepvk/USER-bge-m3", "model-index", "autotrain_compatible", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
sentence-similarity
2025-03-07T12:14:10Z
--- tags: - sentence-transformers - sentence-similarity - feature-extraction - generated_from_trainer - dataset_size:79515 - loss:ContrastiveLoss base_model: deepvk/USER-bge-m3 widget: - source_sentence: Яблоко (f49) IgE, ImmunoCAP sentences: - 'Яблоко, IgE, аллерген - f49. Метод: ImmunoCAP' - 'Фундук, IgE, аллерген - f17. Метод: ИФА' - 'Берёза, (rBet v1 PR-10), IgE, компонент аллергена - t215. Метод: ImmunoCAP' - source_sentence: Молекулярно-генетическое исследование абортивного материала при неразвивающейся беременности sentences: - Латентная железосвязывающая способность сыворотки - Бета-2-микроглобулин в моче - Генетические причины мужского бесплодия. Расширенный - source_sentence: Рентгенография шейно-дорсального и пояснично-крестцового отдела позвоночника sentences: - Рентгеноденситометрия проксимального отдела обеих бедренных костей и поясничного отдела позвоночника - 'Пекарские дрожжи, IgG, аллерген - f45. Метод: ИФА' - 'Вирус Эпштейна-Барра (Epstein-Barr virus): Антитела: IgG, (количественно). Метод: иммуноблот' - source_sentence: Прием (осмотр, консультация) врача-мануального терапевта первичный sentences: - Прием (осмотр, консультация) врача-профпатолога в клинике - Исследование конвергенции - 'Краб, IgE, аллерген - f23. Метод: ИФА' - source_sentence: Определение клиренса эндогенного креатинина sentences: - Консультация врача, в клинике, эндокринолог - Витамин D, 25-гидрокси (кальциферол) - Прием (осмотр, консультация) врача-педиатра дистанционно pipeline_tag: sentence-similarity library_name: sentence-transformers metrics: - cosine_accuracy - cosine_accuracy_threshold - cosine_f1 - cosine_f1_threshold - cosine_precision - cosine_recall - cosine_ap - cosine_mcc model-index: - name: SentenceTransformer based on deepvk/USER-bge-m3 results: - task: type: binary-classification name: Binary Classification dataset: name: binary eval test type: binary-eval-test metrics: - type: cosine_accuracy value: 0.9546068075117371 name: Cosine Accuracy - type: cosine_accuracy_threshold value: 0.7871222496032715 name: Cosine Accuracy Threshold - type: cosine_f1 value: 0.8863058481656375 name: Cosine F1 - type: cosine_f1_threshold value: 0.751617431640625 name: Cosine F1 Threshold - type: cosine_precision value: 0.8778241473593322 name: Cosine Precision - type: cosine_recall value: 0.8949530516431925 name: Cosine Recall - type: cosine_ap value: 0.9361081843605553 name: Cosine Ap - type: cosine_mcc value: 0.857601042471043 name: Cosine Mcc --- # SentenceTransformer based on deepvk/USER-bge-m3 This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [deepvk/USER-bge-m3](https://huggingface.co/deepvk/USER-bge-m3). It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more. ## Model Details ### Model Description - **Model Type:** Sentence Transformer - **Base model:** [deepvk/USER-bge-m3](https://huggingface.co/deepvk/USER-bge-m3) <!-- at revision 0cc6cfe48e260fb0474c753087a69369e88709ae --> - **Maximum Sequence Length:** 8192 tokens - **Output Dimensionality:** 1024 dimensions - **Similarity Function:** Cosine Similarity <!-- - **Training Dataset:** Unknown --> <!-- - **Language:** Unknown --> <!-- - **License:** Unknown --> ### Model Sources - **Documentation:** [Sentence Transformers Documentation](https://sbert.net) - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers) - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers) ### Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 8192, 'do_lower_case': False}) with Transformer model: XLMRobertaModel (1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True}) (2): Normalize() ) ``` ## Usage ### Direct Usage (Sentence Transformers) First install the Sentence Transformers library: ```bash pip install -U sentence-transformers ``` Then you can load this model and run inference. ```python from sentence_transformers import SentenceTransformer # Download from the 🤗 Hub model = SentenceTransformer("EvgenyBondarenko/BiEncoderRanker") # Run inference sentences = [ 'Определение клиренса эндогенного креатинина', 'Консультация врача, в клинике, эндокринолог', 'Прием (осмотр, консультация) врача-педиатра дистанционно', ] embeddings = model.encode(sentences) print(embeddings.shape) # [3, 1024] # Get the similarity scores for the embeddings similarities = model.similarity(embeddings, embeddings) print(similarities.shape) # [3, 3] ``` <!-- ### Direct Usage (Transformers) <details><summary>Click to see the direct usage in Transformers</summary> </details> --> <!-- ### Downstream Usage (Sentence Transformers) You can finetune this model on your own dataset. <details><summary>Click to expand</summary> </details> --> <!-- ### Out-of-Scope Use *List how the model may foreseeably be misused and address what users ought not to do with the model.* --> ## Evaluation ### Metrics #### Binary Classification * Dataset: `binary-eval-test` * Evaluated with [<code>BinaryClassificationEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.BinaryClassificationEvaluator) | Metric | Value | |:--------------------------|:-----------| | cosine_accuracy | 0.9546 | | cosine_accuracy_threshold | 0.7871 | | cosine_f1 | 0.8863 | | cosine_f1_threshold | 0.7516 | | cosine_precision | 0.8778 | | cosine_recall | 0.895 | | **cosine_ap** | **0.9361** | | cosine_mcc | 0.8576 | <!-- ## Bias, Risks and Limitations *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.* --> <!-- ### Recommendations *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.* --> ## Training Details ### Training Dataset #### Unnamed Dataset * Size: 79,515 training samples * Columns: <code>sentence1</code>, <code>sentence2</code>, and <code>label</code> * Approximate statistics based on the first 1000 samples: | | sentence1 | sentence2 | label | |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:------------------------------------------------| | type | string | string | int | | details | <ul><li>min: 4 tokens</li><li>mean: 22.41 tokens</li><li>max: 157 tokens</li></ul> | <ul><li>min: 4 tokens</li><li>mean: 23.01 tokens</li><li>max: 140 tokens</li></ul> | <ul><li>0: ~80.00%</li><li>1: ~20.00%</li></ul> | * Samples: | sentence1 | sentence2 | label | |:------------------------------------------------------------------------------------|:-----------------------------------------------|:---------------| | <code>Тироксин общий (T4 общий, тетрайодтиронин общий, Total Thyroxine, TT4)</code> | <code>Тироксин общий (Т4)</code> | <code>1</code> | | <code>Тироксин общий (T4 общий, тетрайодтиронин общий, Total Thyroxine, TT4)</code> | <code>Трийодтиронин общий (Т3)</code> | <code>0</code> | | <code>Тироксин общий (T4 общий, тетрайодтиронин общий, Total Thyroxine, TT4)</code> | <code>Тироксин свободный (Т4 свободный)</code> | <code>0</code> | * Loss: [<code>ContrastiveLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#contrastiveloss) with these parameters: ```json { "distance_metric": "SiameseDistanceMetric.COSINE_DISTANCE", "margin": 0.5, "size_average": true } ``` ### Evaluation Dataset #### Unnamed Dataset * Size: 34,080 evaluation samples * Columns: <code>sentence1</code>, <code>sentence2</code>, and <code>label</code> * Approximate statistics based on the first 1000 samples: | | sentence1 | sentence2 | label | |:--------|:----------------------------------------------------------------------------------|:----------------------------------------------------------------------------------|:------------------------------------------------| | type | string | string | int | | details | <ul><li>min: 5 tokens</li><li>mean: 21.93 tokens</li><li>max: 78 tokens</li></ul> | <ul><li>min: 4 tokens</li><li>mean: 22.57 tokens</li><li>max: 90 tokens</li></ul> | <ul><li>0: ~80.00%</li><li>1: ~20.00%</li></ul> | * Samples: | sentence1 | sentence2 | label | |:--------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------|:---------------| | <code>Цитологическое исследование соскоба с шейки матки (экзоцервикс,1 стекло)</code> | <code>Жидкостная цитология. Исследование соскоба шейки матки и цервикального канала (окрашивание по Папаниколау)</code> | <code>1</code> | | <code>Цитологическое исследование соскоба с шейки матки (экзоцервикс,1 стекло)</code> | <code>Цитологическое исследование аспирата из полости матки</code> | <code>0</code> | | <code>Цитологическое исследование соскоба с шейки матки (экзоцервикс,1 стекло)</code> | <code>Цитологическое исследование соскобов молочной железы</code> | <code>0</code> | * Loss: [<code>ContrastiveLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#contrastiveloss) with these parameters: ```json { "distance_metric": "SiameseDistanceMetric.COSINE_DISTANCE", "margin": 0.5, "size_average": true } ``` ### Training Hyperparameters #### Non-Default Hyperparameters - `eval_strategy`: steps - `per_device_train_batch_size`: 32 - `per_device_eval_batch_size`: 32 - `learning_rate`: 2e-05 - `num_train_epochs`: 2 - `warmup_ratio`: 0.1 - `save_only_model`: True - `fp16`: True - `load_best_model_at_end`: True #### All Hyperparameters <details><summary>Click to expand</summary> - `overwrite_output_dir`: False - `do_predict`: False - `eval_strategy`: steps - `prediction_loss_only`: True - `per_device_train_batch_size`: 32 - `per_device_eval_batch_size`: 32 - `per_gpu_train_batch_size`: None - `per_gpu_eval_batch_size`: None - `gradient_accumulation_steps`: 1 - `eval_accumulation_steps`: None - `torch_empty_cache_steps`: None - `learning_rate`: 2e-05 - `weight_decay`: 0.0 - `adam_beta1`: 0.9 - `adam_beta2`: 0.999 - `adam_epsilon`: 1e-08 - `max_grad_norm`: 1.0 - `num_train_epochs`: 2 - `max_steps`: -1 - `lr_scheduler_type`: linear - `lr_scheduler_kwargs`: {} - `warmup_ratio`: 0.1 - `warmup_steps`: 0 - `log_level`: passive - `log_level_replica`: warning - `log_on_each_node`: True - `logging_nan_inf_filter`: True - `save_safetensors`: True - `save_on_each_node`: False - `save_only_model`: True - `restore_callback_states_from_checkpoint`: False - `no_cuda`: False - `use_cpu`: False - `use_mps_device`: False - `seed`: 42 - `data_seed`: None - `jit_mode_eval`: False - `use_ipex`: False - `bf16`: False - `fp16`: True - `fp16_opt_level`: O1 - `half_precision_backend`: auto - `bf16_full_eval`: False - `fp16_full_eval`: False - `tf32`: None - `local_rank`: 0 - `ddp_backend`: None - `tpu_num_cores`: None - `tpu_metrics_debug`: False - `debug`: [] - `dataloader_drop_last`: False - `dataloader_num_workers`: 0 - `dataloader_prefetch_factor`: None - `past_index`: -1 - `disable_tqdm`: False - `remove_unused_columns`: True - `label_names`: None - `load_best_model_at_end`: True - `ignore_data_skip`: False - `fsdp`: [] - `fsdp_min_num_params`: 0 - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False} - `fsdp_transformer_layer_cls_to_wrap`: None - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None} - `deepspeed`: None - `label_smoothing_factor`: 0.0 - `optim`: adamw_torch - `optim_args`: None - `adafactor`: False - `group_by_length`: False - `length_column_name`: length - `ddp_find_unused_parameters`: None - `ddp_bucket_cap_mb`: None - `ddp_broadcast_buffers`: False - `dataloader_pin_memory`: True - `dataloader_persistent_workers`: False - `skip_memory_metrics`: True - `use_legacy_prediction_loop`: False - `push_to_hub`: False - `resume_from_checkpoint`: None - `hub_model_id`: None - `hub_strategy`: every_save - `hub_private_repo`: None - `hub_always_push`: False - `gradient_checkpointing`: False - `gradient_checkpointing_kwargs`: None - `include_inputs_for_metrics`: False - `include_for_metrics`: [] - `eval_do_concat_batches`: True - `fp16_backend`: auto - `push_to_hub_model_id`: None - `push_to_hub_organization`: None - `mp_parameters`: - `auto_find_batch_size`: False - `full_determinism`: False - `torchdynamo`: None - `ray_scope`: last - `ddp_timeout`: 1800 - `torch_compile`: False - `torch_compile_backend`: None - `torch_compile_mode`: None - `dispatch_batches`: None - `split_batches`: None - `include_tokens_per_second`: False - `include_num_input_tokens_seen`: False - `neftune_noise_alpha`: None - `optim_target_modules`: None - `batch_eval_metrics`: False - `eval_on_start`: False - `use_liger_kernel`: False - `eval_use_gather_object`: False - `average_tokens_across_devices`: False - `prompts`: None - `batch_sampler`: batch_sampler - `multi_dataset_batch_sampler`: proportional </details> ### Training Logs | Epoch | Step | Training Loss | Validation Loss | binary-eval-test_cosine_ap | |:------:|:----:|:-------------:|:---------------:|:--------------------------:| | 0.2012 | 500 | 0.0131 | 0.0109 | 0.8309 | | 0.4024 | 1000 | 0.01 | 0.0087 | 0.8820 | | 0.6036 | 1500 | 0.0088 | 0.0084 | 0.8865 | | 0.8048 | 2000 | 0.0077 | 0.0069 | 0.9091 | | 1.0060 | 2500 | 0.0073 | 0.0061 | 0.9184 | | 1.2072 | 3000 | 0.0059 | 0.0058 | 0.9261 | | 1.4085 | 3500 | 0.0055 | 0.0057 | 0.9300 | | 1.6097 | 4000 | 0.0054 | 0.0054 | 0.9328 | | 1.8109 | 4500 | 0.0051 | 0.0052 | 0.9361 | ### Framework Versions - Python: 3.10.12 - Sentence Transformers: 3.4.1 - Transformers: 4.49.0 - PyTorch: 2.6.0+cu118 - Accelerate: 1.4.0 - Datasets: 3.3.2 - Tokenizers: 0.21.0 ## Citation ### BibTeX #### Sentence Transformers ```bibtex @inproceedings{reimers-2019-sentence-bert, title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks", author = "Reimers, Nils and Gurevych, Iryna", booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing", month = "11", year = "2019", publisher = "Association for Computational Linguistics", url = "https://arxiv.org/abs/1908.10084", } ``` #### ContrastiveLoss ```bibtex @inproceedings{hadsell2006dimensionality, author={Hadsell, R. and Chopra, S. and LeCun, Y.}, booktitle={2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06)}, title={Dimensionality Reduction by Learning an Invariant Mapping}, year={2006}, volume={2}, number={}, pages={1735-1742}, doi={10.1109/CVPR.2006.100} } ``` <!-- ## Glossary *Clearly define terms in order to be accessible across audiences.* --> <!-- ## Model Card Authors *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.* --> <!-- ## Model Card Contact *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.* -->
dashtoon/hunyuan-video-keyframe-control-lora
dashtoon
2025-03-07T12:15:13Z
0
54
diffusers
[ "diffusers", "base_model:tencent/HunyuanVideo", "base_model:finetune:tencent/HunyuanVideo", "region:us" ]
null
2025-02-24T08:59:05Z
--- base_model: - tencent/HunyuanVideo library_name: diffusers --- HunyuanVideo Keyframe Control Lora is an adapter for HunyuanVideo T2V model for keyframe-based video generation. ​Our architecture builds upon existing models, introducing key enhancements to optimize keyframe-based video generation:​ * We modify the input patch embedding projection layer to effectively incorporate keyframe information. By adjusting the convolutional input parameters, we enable the model to process image inputs within the Diffusion Transformer (DiT) framework.​ * We apply Low-Rank Adaptation (LoRA) across all linear layers and the convolutional input layer. This approach facilitates efficient fine-tuning by introducing low-rank matrices that approximate the weight updates, thereby preserving the base model's foundational capabilities while reducing the number of trainable parameters. * The model is conditioned on user-defined keyframes, allowing precise control over the generated video's start and end frames. This conditioning ensures that the generated content aligns seamlessly with the specified keyframes, enhancing the coherence and narrative flow of the video.​ | Image 1 | Image 2 | Generated Video | |---------|---------|-----------------| | ![Image 1](https://content.dashtoon.ai/stability-images/41aeca63-064a-4003-8c8b-bfe2cc80d275.png) | ![Image 2](https://content.dashtoon.ai/stability-images/28956177-3455-4b56-bb6c-73eacef323ca.png) | <video controls autoplay src="https://content.dashtoon.ai/stability-images/14b7dd1a-1f46-4c4c-b4ec-9d0f948712af.mp4"></video> | | ![Image 1](https://content.dashtoon.ai/stability-images/ddabbf2f-4218-497b-8239-b7b882d93000.png) | ![Image 2](https://content.dashtoon.ai/stability-images/b603acba-40a4-44ba-aa26-ed79403df580.png) | <video controls autoplay src="https://content.dashtoon.ai/stability-images/b00ba193-b3b7-41a1-9bc1-9fdaceba6efa.mp4"></video> | | ![Image 1](https://content.dashtoon.ai/stability-images/5298cf0c-0955-4568-935a-2fb66045f21d.png) | ![Image 2](https://content.dashtoon.ai/stability-images/722a4ea7-7092-4323-8e83-3f627e8fd7f8.png) | <video controls autoplay src="https://content.dashtoon.ai/stability-images/0cb84780-4fdf-4ecc-ab48-12e7e1055a39.mp4"></video> | | ![Image 1](https://content.dashtoon.ai/stability-images/69d9a49f-95c0-4e85-bd49-14a039373c8b.png) | ![Image 2](https://content.dashtoon.ai/stability-images/0cef7fa9-e15a-48ec-9bd3-c61921181802.png) | <video controls autoplay src="https://content.dashtoon.ai/stability-images/ce12156f-0ac2-4d16-b489-37e85c61b5b2.mp4"></video> | ## Code: The tranining code can be found [here](https://github.com/dashtoon/hunyuan-video-keyframe-control-lora). ## Recommended Settings 1. The model works best on human subjects. Single subject images work slightly better. 2. It is recommended to use the following image generation resolutions `720x1280`, `544x960`, `1280x720`, `960x544`. 3. It is recommended to set frames from 33 upto 97. Can go upto 121 frames as well (but not tested much). 4. Prompting helps a lot but works even without. The prompt can be as simple as just the name of the object you want to generate or can be detailed. 5. `num_inference_steps` is recommended to be 50, but for fast results you can use 30 as well. Anything less than 30 is not recommended. ## Diffusers HunyuanVideo Keyframe Control Lora can be used directly from Diffusers. Install the latest version of Diffusers. ## Inference While the included `inference.py` script can be used to run inference. We would encourage folks to visit out [github repo](https://github.com/dashtoon/hunyuan-video-keyframe-control-lora/blob/main/hv_control_lora_inference.py) which contains a much optimized version of this inference script.
techdkit/mumbai-call-girls-7290901024
techdkit
2025-03-07T12:13:09Z
0
0
null
[ "region:us" ]
null
2025-03-07T12:12:58Z
<p>Looking for the&nbsp;<strong>best Russian escorts in Chandigarh</strong>? Experience ultimate pleasure and companionship with our&nbsp;<strong>high-class call girls</strong>, available in&nbsp;<strong>Aerocity,&nbsp;Chandigarh, and luxury hotels across Chandigarh</strong>. Whether you seek a passionate night, an elegant date, or a sensual massage, our&nbsp;<strong>premium Russian escorts</strong>&nbsp;ensure an unforgettable experience.&nbsp;100%&nbsp;<strong>discreet &amp; safe services</strong></p> <h3><strong><a href="https://api.whatsapp.com/send/?phone=917290901024&amp;text&amp;type=phone_number&amp;app_absent=0">WhatsApp Me: 7290901024</a></strong></h3> <p><strong>Visit Website &gt; <a href="https://mumbai.noida-escort.com/">https://mumbai.noida-escort.com</a></strong></p> <p>Call Girls girl in Chandigarh&nbsp;call girls Call Girl service in Chandigarh&nbsp;Chandigarh Call Girls Locanto&nbsp;call girls Call Girl service in Chandigarh&nbsp;Gurgaonsex Call Girls&nbsp;Escorts Service at Chandigarh&nbsp;Chandigarh Escort Service&nbsp;Chandigarh Escort girl&nbsp;Escort service Chandigarh&nbsp;female Escort in Chandigarh&nbsp;Gurgaonsex girl&nbsp;Call Girls in Chandigarh&nbsp;Call Girls services Chandigarh&nbsp;Chandigarh female Escorts&nbsp;housewife Escort in Chandigarh&nbsp;Chandigarh Escort&nbsp;Chandigarh Escorts&nbsp;Chandigarh Escort service&nbsp;model Escorts in Chandigarh&nbsp;Chandigarh Call Girls services&nbsp;Call Girls service Chandigarh&nbsp;Russian Escort in Chandigarh&nbsp;call girls near me&nbsp;Chandigarh Escoorts Call Girls&nbsp;royal Call Girls in Chandigarh&nbsp;Chandigarh independent Escort&nbsp;cheap Chandigarh Escort&nbsp;Escort girls Service Chandigarh&nbsp;independent Chandigarh Call Girls&nbsp;foreign Escort Girls in Chandigarh&nbsp;cheap female Escort Girls in Chandigarh&nbsp;royal Call Girls Chandigarh&nbsp;hotel Call Girls in Chandigarh&nbsp;Chandigarh Call Girls girls&nbsp;female Call Girl Chandigarh&nbsp;sex Call Girls in Chandigarh&nbsp;Call Girl service near me&nbsp;call girls&nbsp;Russian Call Girls Chandigarh&nbsp;independent Call Girls in Chandigarh&nbsp;cheap Call Girls in Chandigarh&nbsp;model Call Girls Chandigarh&nbsp;Call Girls service in Chandigarh&nbsp;Call Girls Chandigarh&nbsp;female Call Girls Chandigarh&nbsp;Call Girls services in Chandigarh&nbsp;GurgaonVIP Call Girls&nbsp;call girl near me&nbsp;Gurgaonsex&nbsp;Gurgaoncall girl number&nbsp;independent Call Girls Chandigarh&nbsp;cheap Call Girls Chandigarh&nbsp;girls Call Girl Chandigarh&nbsp;call girls Call Girls service in Chandigarh&nbsp;Call Girls service in Chandigarh&nbsp;independent call girls&nbsp;VIP Call Girls in Chandigarh&nbsp;Gurgaonmodel Call Girls&nbsp;female Call Girls of Chandigarh&nbsp;Chandigarh Call Girl agency&nbsp;Chandigarh Call Girl service&nbsp;call girls Call Girl service in Chandigarh&nbsp;GurgaonRussian Call Girls&nbsp;Gurgaoni sex video&nbsp;call girl number&nbsp;call girls Chandigarh&nbsp;Chandigarh Call Girls agency&nbsp;call girls Call Girl service in Chandigarh&nbsp;Gurgaonsex&nbsp;call girl contact number&nbsp;sexy Call Girls call girls in Chandigarh&nbsp;Chandigarh Call Girls&nbsp;कॉल गर्ल लिस्ट Chandigarh&nbsp;call girls Call Girl service in Chandigarh&nbsp;call girl in Chandigarh&nbsp;best Call Girls in Chandigarh&nbsp;Call Girls Chandigarh</p>
robiulawaldev/9be280cd-89e2-4e4e-8f72-ad4d05886f00
robiulawaldev
2025-03-07T12:12:15Z
0
0
peft
[ "peft", "generated_from_trainer", "base_model:codellama/CodeLlama-7b-Instruct-hf", "base_model:adapter:codellama/CodeLlama-7b-Instruct-hf", "region:us" ]
null
2025-03-07T12:11:57Z
--- library_name: peft tags: - generated_from_trainer base_model: codellama/CodeLlama-7b-Instruct-hf model-index: - name: robiulawaldev/9be280cd-89e2-4e4e-8f72-ad4d05886f00 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # robiulawaldev/9be280cd-89e2-4e4e-8f72-ad4d05886f00 This model was trained from scratch on the None dataset. It achieves the following results on the evaluation set: - Loss: nan ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ### Framework versions - PEFT 0.13.2 - Transformers 4.46.0 - Pytorch 2.5.0+cu124 - Datasets 3.0.1 - Tokenizers 0.20.1
techdkit/chandigarh-call-girls-service7290901024
techdkit
2025-03-07T12:11:44Z
0
0
null
[ "region:us" ]
null
2025-03-07T12:10:53Z
<p>Looking for the&nbsp;<strong>best Russian escorts in Chandigarh</strong>? Experience ultimate pleasure and companionship with our&nbsp;<strong>high-class call girls</strong>, available in&nbsp;<strong>Aerocity,&nbsp;Chandigarh, and luxury hotels across Chandigarh</strong>. Whether you seek a passionate night, an elegant date, or a sensual massage, our&nbsp;<strong>premium Russian escorts</strong>&nbsp;ensure an unforgettable experience.&nbsp;100%&nbsp;<strong>discreet &amp; safe services</strong></p> <h3><strong><a href="https://api.whatsapp.com/send/?phone=917290901024&amp;text&amp;type=phone_number&amp;app_absent=0">WhatsApp Me: 7290901024</a></strong></h3> <p><strong>Visit Website &gt; <a href="https://chandigarh.noida-escort.com/">https://chandigarh.noida-escort.com</a></strong></p> <p>Call Girls girl in Chandigarh&nbsp;call girls Call Girl service in Chandigarh&nbsp;Chandigarh Call Girls Locanto&nbsp;call girls Call Girl service in Chandigarh&nbsp;Gurgaonsex Call Girls&nbsp;Escorts Service at Chandigarh&nbsp;Chandigarh Escort Service&nbsp;Chandigarh Escort girl&nbsp;Escort service Chandigarh&nbsp;female Escort in Chandigarh&nbsp;Gurgaonsex girl&nbsp;Call Girls in Chandigarh&nbsp;Call Girls services Chandigarh&nbsp;Chandigarh female Escorts&nbsp;housewife Escort in Chandigarh&nbsp;Chandigarh Escort&nbsp;Chandigarh Escorts&nbsp;Chandigarh Escort service&nbsp;model Escorts in Chandigarh&nbsp;Chandigarh Call Girls services&nbsp;Call Girls service Chandigarh&nbsp;Russian Escort in Chandigarh&nbsp;call girls near me&nbsp;Chandigarh Escoorts Call Girls&nbsp;royal Call Girls in Chandigarh&nbsp;Chandigarh independent Escort&nbsp;cheap Chandigarh Escort&nbsp;Escort girls Service Chandigarh&nbsp;independent Chandigarh Call Girls&nbsp;foreign Escort Girls in Chandigarh&nbsp;cheap female Escort Girls in Chandigarh&nbsp;royal Call Girls Chandigarh&nbsp;hotel Call Girls in Chandigarh&nbsp;Chandigarh Call Girls girls&nbsp;female Call Girl Chandigarh&nbsp;sex Call Girls in Chandigarh&nbsp;Call Girl service near me&nbsp;call girls&nbsp;Russian Call Girls Chandigarh&nbsp;independent Call Girls in Chandigarh&nbsp;cheap Call Girls in Chandigarh&nbsp;model Call Girls Chandigarh&nbsp;Call Girls service in Chandigarh&nbsp;Call Girls Chandigarh&nbsp;female Call Girls Chandigarh&nbsp;Call Girls services in Chandigarh&nbsp;GurgaonVIP Call Girls&nbsp;call girl near me&nbsp;Gurgaonsex&nbsp;Gurgaoncall girl number&nbsp;independent Call Girls Chandigarh&nbsp;cheap Call Girls Chandigarh&nbsp;girls Call Girl Chandigarh&nbsp;call girls Call Girls service in Chandigarh&nbsp;Call Girls service in Chandigarh&nbsp;independent call girls&nbsp;VIP Call Girls in Chandigarh&nbsp;Gurgaonmodel Call Girls&nbsp;female Call Girls of Chandigarh&nbsp;Chandigarh Call Girl agency&nbsp;Chandigarh Call Girl service&nbsp;call girls Call Girl service in Chandigarh&nbsp;GurgaonRussian Call Girls&nbsp;Gurgaoni sex video&nbsp;call girl number&nbsp;call girls Chandigarh&nbsp;Chandigarh Call Girls agency&nbsp;call girls Call Girl service in Chandigarh&nbsp;Gurgaonsex&nbsp;call girl contact number&nbsp;sexy Call Girls call girls in Chandigarh&nbsp;Chandigarh Call Girls&nbsp;कॉल गर्ल लिस्ट Chandigarh&nbsp;call girls Call Girl service in Chandigarh&nbsp;call girl in Chandigarh&nbsp;best Call Girls in Chandigarh&nbsp;Call Girls Chandigarh</p>
mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF
mradermacher
2025-03-07T12:10:56Z
0
0
transformers
[ "transformers", "gguf", "en", "dataset:BAAI/Infinity-Instruct", "base_model:BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference", "base_model:quantized:BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference", "endpoints_compatible", "region:us", "imatrix", "conversational" ]
null
2025-03-07T11:24:06Z
--- base_model: BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference datasets: - BAAI/Infinity-Instruct language: - en library_name: transformers quantized_by: mradermacher --- ## About <!-- ### quantize_version: 2 --> <!-- ### output_tensor_quantised: 1 --> <!-- ### convert_type: hf --> <!-- ### vocab_type: --> <!-- ### tags: nicoboss --> weighted/imatrix quants of https://huggingface.co/BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference <!-- provided-files --> static quants are available at https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-GGUF ## Usage If you are unsure how to use GGUF files, refer to one of [TheBloke's READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for more details, including on how to concatenate multi-part files. ## Provided Quants (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-IQ1_S.gguf) | i1-IQ1_S | 2.5 | for the desperate | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-IQ1_M.gguf) | i1-IQ1_M | 2.6 | mostly desperate | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 2.9 | | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-IQ2_XS.gguf) | i1-IQ2_XS | 3.2 | | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-IQ2_S.gguf) | i1-IQ2_S | 3.3 | | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-IQ2_M.gguf) | i1-IQ2_M | 3.5 | | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-Q2_K_S.gguf) | i1-Q2_K_S | 3.7 | very low quality | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 3.9 | lower quality | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-Q2_K.gguf) | i1-Q2_K | 3.9 | IQ3_XXS probably better | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-IQ3_XS.gguf) | i1-IQ3_XS | 4.2 | | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-IQ3_S.gguf) | i1-IQ3_S | 4.4 | beats Q3_K* | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-Q3_K_S.gguf) | i1-Q3_K_S | 4.4 | IQ3_XS probably better | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-IQ3_M.gguf) | i1-IQ3_M | 4.6 | | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-Q3_K_M.gguf) | i1-Q3_K_M | 4.9 | IQ3_S probably better | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-Q3_K_L.gguf) | i1-Q3_K_L | 5.2 | IQ3_M probably better | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-IQ4_XS.gguf) | i1-IQ4_XS | 5.3 | | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-IQ4_NL.gguf) | i1-IQ4_NL | 5.5 | prefer IQ4_XS | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-Q4_0.gguf) | i1-Q4_0 | 5.6 | fast, low quality | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-Q4_K_S.gguf) | i1-Q4_K_S | 5.6 | optimal size/speed/quality | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-Q4_K_M.gguf) | i1-Q4_K_M | 5.9 | fast, recommended | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-Q4_1.gguf) | i1-Q4_1 | 6.1 | | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-Q5_K_S.gguf) | i1-Q5_K_S | 6.6 | | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-Q5_K_M.gguf) | i1-Q5_K_M | 6.7 | | | [GGUF](https://huggingface.co/mradermacher/Gemma2-9B-IT-Simpo-Infinity-Preference-i1-GGUF/resolve/main/Gemma2-9B-IT-Simpo-Infinity-Preference.i1-Q6_K.gguf) | i1-Q6_K | 7.7 | practically like static Q6_K | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): ![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 ## FAQ / Model Request See https://huggingface.co/mradermacher/model_requests for some answers to questions you might have and/or if you want some other model quantized. ## Thanks I thank my company, [nethype GmbH](https://www.nethype.de/), for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. <!-- end -->
Laocaicai/DeepSeek-R1-Distill-Qwen-7B-Q6_K-GGUF
Laocaicai
2025-03-07T12:10:03Z
0
0
transformers
[ "transformers", "gguf", "llama-cpp", "gguf-my-repo", "base_model:deepseek-ai/DeepSeek-R1-Distill-Qwen-7B", "base_model:quantized:deepseek-ai/DeepSeek-R1-Distill-Qwen-7B", "license:mit", "endpoints_compatible", "region:us", "conversational" ]
null
2025-03-07T12:09:35Z
--- base_model: deepseek-ai/DeepSeek-R1-Distill-Qwen-7B library_name: transformers license: mit tags: - llama-cpp - gguf-my-repo --- # Laocaicai/DeepSeek-R1-Distill-Qwen-7B-Q6_K-GGUF This model was converted to GGUF format from [`deepseek-ai/DeepSeek-R1-Distill-Qwen-7B`](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-7B) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space. Refer to the [original model card](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-7B) for more details on the model. ## Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) ```bash brew install llama.cpp ``` Invoke the llama.cpp server or the CLI. ### CLI: ```bash llama-cli --hf-repo Laocaicai/DeepSeek-R1-Distill-Qwen-7B-Q6_K-GGUF --hf-file deepseek-r1-distill-qwen-7b-q6_k.gguf -p "The meaning to life and the universe is" ``` ### Server: ```bash llama-server --hf-repo Laocaicai/DeepSeek-R1-Distill-Qwen-7B-Q6_K-GGUF --hf-file deepseek-r1-distill-qwen-7b-q6_k.gguf -c 2048 ``` Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well. Step 1: Clone llama.cpp from GitHub. ``` git clone https://github.com/ggerganov/llama.cpp ``` Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux). ``` cd llama.cpp && LLAMA_CURL=1 make ``` Step 3: Run inference through the main binary. ``` ./llama-cli --hf-repo Laocaicai/DeepSeek-R1-Distill-Qwen-7B-Q6_K-GGUF --hf-file deepseek-r1-distill-qwen-7b-q6_k.gguf -p "The meaning to life and the universe is" ``` or ``` ./llama-server --hf-repo Laocaicai/DeepSeek-R1-Distill-Qwen-7B-Q6_K-GGUF --hf-file deepseek-r1-distill-qwen-7b-q6_k.gguf -c 2048 ```
MrRobotoAI/D13
MrRobotoAI
2025-03-07T12:09:10Z
20
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "mergekit", "merge", "conversational", "arxiv:2203.05482", "base_model:MrRobotoAI/D11", "base_model:merge:MrRobotoAI/D11", "base_model:MrRobotoAI/D6", "base_model:merge:MrRobotoAI/D6", "base_model:MrRobotoAI/L2", "base_model:merge:MrRobotoAI/L2", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-06T20:15:00Z
--- base_model: - MrRobotoAI/137 - MrRobotoAI/135 - MrRobotoAI/134 - MrRobotoAI/133 - MrRobotoAI/138 - MrRobotoAI/136 - MrRobotoAI/L2 library_name: transformers tags: - mergekit - merge --- # merge 13,027 This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [Linear](https://arxiv.org/abs/2203.05482) merge method. ### Models Merged The following models were included in the merge: * [MrRobotoAI/137](https://huggingface.co/MrRobotoAI/137) * [MrRobotoAI/135](https://huggingface.co/MrRobotoAI/135) * [MrRobotoAI/134](https://huggingface.co/MrRobotoAI/134) * [MrRobotoAI/133](https://huggingface.co/MrRobotoAI/133) * [MrRobotoAI/138](https://huggingface.co/MrRobotoAI/138) * [MrRobotoAI/136](https://huggingface.co/MrRobotoAI/136) * [MrRobotoAI/L2](https://huggingface.co/MrRobotoAI/L2) ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: MrRobotoAI/133 - model: MrRobotoAI/134 - model: MrRobotoAI/135 - model: MrRobotoAI/136 - model: MrRobotoAI/137 - model: MrRobotoAI/138 - model: MrRobotoAI/L2 parameters: weight: 1.0 merge_method: linear dtype: float16 ```
AQSlim58/AQSlim
AQSlim58
2025-03-07T12:09:05Z
0
0
null
[ "region:us" ]
null
2025-03-07T12:07:58Z
# AQ Slim Nederland - Ervaring Prijs, Kopen AQ Slim is een innovatief, natuurlijk supplement dat je helpt bij het behalen van je gewichtsverliesdoelen, het versnellen van je stofwisseling en het verhogen van je energieniveau. Met een zorgvuldig samengestelde mix van krachtige ingrediënten zoals groene thee-extract, cafeïne en L-carnitine, werkt AQ Slim samen met je lichaam om vetverbranding te stimuleren, je metabolisme te versnellen en je energie te boosten, zonder schadelijke kunstmatige toevoegingen. ## **[Klik hier om te bestellen op de officiële website van AQ Slim](https://gezondekliniek.nl/product/aq-slim/)** ## Hoe Werkt AQ Slim? AQ Slim werkt door verschillende mechanismen die het vetverbrandingsproces in je lichaam ondersteunen. Het hoofddoel is om de stofwisseling te versnellen en je lichaam in staat te stellen vet effectiever te verbranden voor energie. Laten we enkele van de belangrijkste werkingsprincipes van AQ Slim bespreken. ### 1. Verhoogde Thermogenese Een van de belangrijkste voordelen van AQ Slim is het vermogen om de thermogenese in je lichaam te verhogen. Thermogenese is het proces waarbij het lichaam warmte produceert door vet te verbranden. Dit helpt niet alleen bij gewichtsverlies, maar verhoogt ook je energieniveau. AQ Slim bevat groene thee-extract en cafeïne, die de vetverbranding helpen versnellen. Deze ingrediënten verhogen je lichaamstemperatuur en zorgen ervoor dat je meer calorieën verbrandt, zelfs wanneer je niet fysiek actief bent. ### 2. Vetverbranding en Energieproductie AQ Slim maakt gebruik van L-carnitine, een stof die het lichaam helpt vet om te zetten in energie. Hierdoor krijgt je lichaam de energie die het nodig heeft om dagelijkse activiteiten en zelfs intensieve trainingen aan te kunnen. L-carnitine speelt een belangrijke rol in het transporteren van vetzuren naar de mitochondriën, de energiecentrales van je cellen, waar ze worden verbrand voor energie. Dit betekent dat je niet alleen vet verliest, maar dat je ook een constante energieboost hebt, wat je helpt om je dagelijkse taken en trainingen te verbeteren. ###3. Verbeterde Stofwisseling Een van de grootste obstakels voor gewichtsverlies is een trage stofwisseling. AQ Slim helpt je stofwisseling te versnellen, waardoor je lichaam efficiënter calorieën verbrandt. Het verhoogt het metabolisme, waardoor je lichaam vet sneller omzet in energie. Dit is vooral belangrijk als je probeert af te vallen, omdat een versnelde stofwisseling betekent dat je meer calorieën verbrandt, zelfs wanneer je in rust bent. ## Wie zou AQ Slim Moeten Gebruiken? AQ Slim is ideaal voor mensen die hun gewichtsverliesdoelen willen versnellen en hun algehele gezondheid willen verbeteren. Het is geschikt voor iedereen die zijn metabolisme wil versnellen, vet wil verbranden en tegelijkertijd zijn energie en focus wil verhogen. Dit supplement is bijzonder effectief voor: Mensen die moeite hebben met het verliezen van gewicht, ondanks een gezond dieet en regelmatige lichaamsbeweging. Iedereen die zijn energieniveaus wil verhogen voor een actieve levensstijl. Mensen die op zoek zijn naar een natuurlijk alternatief voor kunstmatige afslankproducten. Atleten of fitnessliefhebbers die hun prestaties willen verbeteren en vet willen verbranden. ## Conclusie: Waarom AQ Slim een Uitstekende Keuze Is AQ Slim is een uitstekend product voor iedereen die op zoek is naar een veilige, natuurlijke en effectieve manier om gewichtsverlies te ondersteunen. Met zijn krachtige ingrediënten die de vetverbranding bevorderen, het metabolisme versnellen en de energieproductie verhogen, is AQ Slim meer dan alleen een afslankmiddel. Het is een hulpmiddel voor een gezondere levensstijl die je helpt om je doelen te bereiken en te behouden. Of je nu probeert af te vallen, je energie te verbeteren, of gewoon je algehele gezondheid te optimaliseren, AQ Slim biedt de ondersteuning die je nodig hebt. Het is de perfecte aanvulling op een gezonde levensstijl en biedt alles wat je nodig hebt om langdurige en duurzame resultaten te behalen. Begin vandaag nog met AQ Slim en ontdek het verschil in je gezondheid, energie en welzijn! ## **[Klik hier om te bestellen op de officiële website van AQ Slim](https://gezondekliniek.nl/product/aq-slim/)**
deenesh01981/aiautomation
deenesh01981
2025-03-07T12:08:34Z
0
0
null
[ "license:bigscience-openrail-m", "region:us" ]
null
2025-03-07T12:08:34Z
--- license: bigscience-openrail-m ---
techdkit/call-girls-in-Singapore-7290901024
techdkit
2025-03-07T12:07:43Z
0
0
null
[ "region:us" ]
null
2025-03-07T12:06:57Z
<p>Looking for the&nbsp;<strong>best Russian escorts in Singapore</strong>? Experience ultimate pleasure and companionship with our&nbsp;<strong>high-class call girls</strong>, available in&nbsp;<strong>Aerocity,&nbsp;Singapore, and luxury hotels across Singapore</strong>. Whether you seek a passionate night, an elegant date, or a sensual massage, our&nbsp;<strong>premium Russian escorts</strong>&nbsp;ensure an unforgettable experience.&nbsp;100%&nbsp;<strong>discreet &amp; safe services</strong></p> <h3><strong><a href="https://api.whatsapp.com/send/?phone=917290901024&amp;text&amp;type=phone_number&amp;app_absent=0">WhatsApp Me: 7290901024</a></strong></h3> <p><strong>Visit Website &gt; <a href="https://singapore.noida-escort.com/">https://singapore.noida-escort.com</a></strong></p> <p>Call Girls girl in Singapore&nbsp;call girls Call Girl service in Singapore&nbsp;Singapore Call Girls Locanto&nbsp;call girls Call Girl service in Singapore&nbsp;Gurgaonsex Call Girls&nbsp;Escorts Service at Singapore&nbsp;Singapore Escort Service&nbsp;Singapore Escort girl&nbsp;Escort service Singapore&nbsp;female Escort in Singapore&nbsp;Gurgaonsex girl&nbsp;Call Girls in Singapore&nbsp;Call Girls services Singapore&nbsp;Singapore female Escorts&nbsp;housewife Escort in Singapore&nbsp;Singapore Escort&nbsp;Singapore Escorts&nbsp;Singapore Escort service&nbsp;model Escorts in Singapore&nbsp;Singapore Call Girls services&nbsp;Call Girls service Singapore&nbsp;Russian Escort in Singapore&nbsp;call girls near me&nbsp;Singapore Escoorts Call Girls&nbsp;royal Call Girls in Singapore&nbsp;Singapore independent Escort&nbsp;cheap Singapore Escort&nbsp;Escort girls Service Singapore&nbsp;independent Singapore Call Girls&nbsp;foreign Escort Girls in Singapore&nbsp;cheap female Escort Girls in Singapore&nbsp;royal Call Girls Singapore&nbsp;hotel Call Girls in Singapore&nbsp;Singapore Call Girls girls&nbsp;female Call Girl Singapore&nbsp;sex Call Girls in Singapore&nbsp;Call Girl service near me&nbsp;call girls&nbsp;Russian Call Girls Singapore&nbsp;independent Call Girls in Singapore&nbsp;cheap Call Girls in Singapore&nbsp;model Call Girls Singapore&nbsp;Call Girls service in Singapore&nbsp;Call Girls Singapore&nbsp;female Call Girls Singapore&nbsp;Call Girls services in Singapore&nbsp;GurgaonVIP Call Girls&nbsp;call girl near me&nbsp;Gurgaonsex&nbsp;Gurgaoncall girl number&nbsp;independent Call Girls Singapore&nbsp;cheap Call Girls Singapore&nbsp;girls Call Girl Singapore&nbsp;call girls Call Girls service in Singapore&nbsp;Call Girls service in Singapore&nbsp;independent call girls&nbsp;VIP Call Girls in Singapore&nbsp;Gurgaonmodel Call Girls&nbsp;female Call Girls of Singapore&nbsp;Singapore Call Girl agency&nbsp;Singapore Call Girl service&nbsp;call girls Call Girl service in Singapore&nbsp;GurgaonRussian Call Girls&nbsp;Gurgaoni sex video&nbsp;call girl number&nbsp;call girls Singapore&nbsp;Singapore Call Girls agency&nbsp;call girls Call Girl service in Singapore&nbsp;Gurgaonsex&nbsp;call girl contact number&nbsp;sexy Call Girls call girls in Singapore&nbsp;Singapore Call Girls&nbsp;कॉल गर्ल लिस्ट Singapore&nbsp;call girls Call Girl service in Singapore&nbsp;call girl in Singapore&nbsp;best Call Girls in Singapore&nbsp;Call Girls Singapore</p>
nindzhal/gemma-2-2B-it-thinking-function_calling-V0
nindzhal
2025-03-07T12:05:21Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "generated_from_trainer", "trl", "sft", "base_model:google/gemma-2-2b-it", "base_model:finetune:google/gemma-2-2b-it", "endpoints_compatible", "region:us" ]
null
2025-03-07T11:51:01Z
--- base_model: google/gemma-2-2b-it library_name: transformers model_name: gemma-2-2B-it-thinking-function_calling-V0 tags: - generated_from_trainer - trl - sft licence: license --- # Model Card for gemma-2-2B-it-thinking-function_calling-V0 This model is a fine-tuned version of [google/gemma-2-2b-it](https://huggingface.co/google/gemma-2-2b-it). It has been trained using [TRL](https://github.com/huggingface/trl). ## Quick start ```python from transformers import pipeline question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?" generator = pipeline("text-generation", model="nindzhal/gemma-2-2B-it-thinking-function_calling-V0", device="cuda") output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0] print(output["generated_text"]) ``` ## Training procedure This model was trained with SFT. ### Framework versions - TRL: 0.15.2 - Transformers: 4.49.0 - Pytorch: 2.5.1 - Datasets: 3.3.2 - Tokenizers: 0.21.0 ## Citations Cite TRL as: ```bibtex @misc{vonwerra2022trl, title = {{TRL: Transformer Reinforcement Learning}}, author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec}, year = 2020, journal = {GitHub repository}, publisher = {GitHub}, howpublished = {\url{https://github.com/huggingface/trl}} } ```
Laocaicai/Qwen2.5-7B-Instruct-Uncensored-Q6_K-GGUF
Laocaicai
2025-03-07T11:59:20Z
0
0
null
[ "gguf", "qwen", "uncensored", "llama-cpp", "gguf-my-repo", "text-generation", "zh", "en", "dataset:NobodyExistsOnTheInternet/ToxicQAFinal", "dataset:anthracite-org/kalo-opus-instruct-22k-no-refusal", "dataset:Orion-zhen/dpo-toxic-zh", "dataset:unalignment/toxic-dpo-v0.2", "dataset:Crystalcareai/Intel-DPO-Pairs-Norefusals", "base_model:Orion-zhen/Qwen2.5-7B-Instruct-Uncensored", "base_model:quantized:Orion-zhen/Qwen2.5-7B-Instruct-Uncensored", "license:gpl-3.0", "model-index", "endpoints_compatible", "region:us", "conversational" ]
text-generation
2025-03-07T11:58:52Z
--- base_model: Orion-zhen/Qwen2.5-7B-Instruct-Uncensored datasets: - NobodyExistsOnTheInternet/ToxicQAFinal - anthracite-org/kalo-opus-instruct-22k-no-refusal - Orion-zhen/dpo-toxic-zh - unalignment/toxic-dpo-v0.2 - Crystalcareai/Intel-DPO-Pairs-Norefusals language: - zh - en license: gpl-3.0 pipeline_tag: text-generation tags: - qwen - uncensored - llama-cpp - gguf-my-repo model-index: - name: Qwen2.5-7B-Instruct-Uncensored results: - task: type: text-generation name: Text Generation dataset: name: IFEval (0-Shot) type: HuggingFaceH4/ifeval args: num_few_shot: 0 metrics: - type: inst_level_strict_acc and prompt_level_strict_acc value: 72.04 name: strict accuracy source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Orion-zhen/Qwen2.5-7B-Instruct-Uncensored name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: BBH (3-Shot) type: BBH args: num_few_shot: 3 metrics: - type: acc_norm value: 35.83 name: normalized accuracy source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Orion-zhen/Qwen2.5-7B-Instruct-Uncensored name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: MATH Lvl 5 (4-Shot) type: hendrycks/competition_math args: num_few_shot: 4 metrics: - type: exact_match value: 1.36 name: exact match source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Orion-zhen/Qwen2.5-7B-Instruct-Uncensored name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: GPQA (0-shot) type: Idavidrein/gpqa args: num_few_shot: 0 metrics: - type: acc_norm value: 7.05 name: acc_norm source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Orion-zhen/Qwen2.5-7B-Instruct-Uncensored name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: MuSR (0-shot) type: TAUR-Lab/MuSR args: num_few_shot: 0 metrics: - type: acc_norm value: 13.58 name: acc_norm source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Orion-zhen/Qwen2.5-7B-Instruct-Uncensored name: Open LLM Leaderboard - task: type: text-generation name: Text Generation dataset: name: MMLU-PRO (5-shot) type: TIGER-Lab/MMLU-Pro config: main split: test args: num_few_shot: 5 metrics: - type: acc value: 38.07 name: accuracy source: url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Orion-zhen/Qwen2.5-7B-Instruct-Uncensored name: Open LLM Leaderboard --- # Laocaicai/Qwen2.5-7B-Instruct-Uncensored-Q6_K-GGUF This model was converted to GGUF format from [`Orion-zhen/Qwen2.5-7B-Instruct-Uncensored`](https://huggingface.co/Orion-zhen/Qwen2.5-7B-Instruct-Uncensored) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space. Refer to the [original model card](https://huggingface.co/Orion-zhen/Qwen2.5-7B-Instruct-Uncensored) for more details on the model. ## Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) ```bash brew install llama.cpp ``` Invoke the llama.cpp server or the CLI. ### CLI: ```bash llama-cli --hf-repo Laocaicai/Qwen2.5-7B-Instruct-Uncensored-Q6_K-GGUF --hf-file qwen2.5-7b-instruct-uncensored-q6_k.gguf -p "The meaning to life and the universe is" ``` ### Server: ```bash llama-server --hf-repo Laocaicai/Qwen2.5-7B-Instruct-Uncensored-Q6_K-GGUF --hf-file qwen2.5-7b-instruct-uncensored-q6_k.gguf -c 2048 ``` Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well. Step 1: Clone llama.cpp from GitHub. ``` git clone https://github.com/ggerganov/llama.cpp ``` Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux). ``` cd llama.cpp && LLAMA_CURL=1 make ``` Step 3: Run inference through the main binary. ``` ./llama-cli --hf-repo Laocaicai/Qwen2.5-7B-Instruct-Uncensored-Q6_K-GGUF --hf-file qwen2.5-7b-instruct-uncensored-q6_k.gguf -p "The meaning to life and the universe is" ``` or ``` ./llama-server --hf-repo Laocaicai/Qwen2.5-7B-Instruct-Uncensored-Q6_K-GGUF --hf-file qwen2.5-7b-instruct-uncensored-q6_k.gguf -c 2048 ```
techdkit/surat-call-girls-7290901024
techdkit
2025-03-07T11:57:32Z
0
0
null
[ "region:us" ]
null
2025-03-07T11:56:20Z
<p>Looking for the&nbsp;<strong>best Russian escorts in Delhi</strong>? Experience ultimate pleasure and companionship with our&nbsp;<strong>high-class call girls</strong>, available in&nbsp;<strong>Aerocity,&nbsp;Surat, and luxury hotels across Surat</strong>. Whether you seek a passionate night, an elegant date, or a sensual massage, our&nbsp;<strong>premium Russian escorts</strong>&nbsp;ensure an unforgettable experience.&nbsp;100%&nbsp;<strong>discreet &amp; safe services</strong></p> <h3><strong><a href="https://api.whatsapp.com/send/?phone=917290901024&amp;text&amp;type=phone_number&amp;app_absent=0">WhatsApp Me: 7290901024</a></strong></h3> <p><strong>Visit Website &gt; <a href="https://surat.noidacallgirls.in/">https://surat.noidacallgirls.in/</a></strong></p> <p>Call Girls girl in Surat&nbsp;call girls Call Girl service in Surat&nbsp;Surat Call Girls Locanto&nbsp;call girls Call Girl service in Surat&nbsp;Gurgaonsex Call Girls&nbsp;Escorts Service at Surat&nbsp;Surat Escort Service&nbsp;Surat Escort girl&nbsp;Escort service Surat&nbsp;female Escort in Surat&nbsp;Gurgaonsex girl&nbsp;Call Girls in Surat&nbsp;Call Girls services Surat&nbsp;Surat female Escorts&nbsp;housewife Escort in Surat&nbsp;Surat Escort&nbsp;Surat Escorts&nbsp;Surat Escort service&nbsp;model Escorts in Surat&nbsp;Surat Call Girls services&nbsp;Call Girls service Surat&nbsp;Russian Escort in Surat&nbsp;call girls near me&nbsp;Surat Escoorts Call Girls&nbsp;royal Call Girls in Surat&nbsp;Surat independent Escort&nbsp;cheap Surat Escort&nbsp;Escort girls Service Surat&nbsp;independent Surat Call Girls&nbsp;foreign Escort Girls in Surat&nbsp;cheap female Escort Girls in Surat&nbsp;royal Call Girls Surat&nbsp;hotel Call Girls in Surat&nbsp;Surat Call Girls girls&nbsp;female Call Girl Surat&nbsp;sex Call Girls in Surat&nbsp;Call Girl service near me&nbsp;call girls&nbsp;Russian Call Girls Surat&nbsp;independent Call Girls in Surat&nbsp;cheap Call Girls in Surat&nbsp;model Call Girls Surat&nbsp;Call Girls service in Surat&nbsp;Call Girls Surat&nbsp;female Call Girls Surat&nbsp;Call Girls services in Surat&nbsp;GurgaonVIP Call Girls&nbsp;call girl near me&nbsp;Gurgaonsex&nbsp;Gurgaoncall girl number&nbsp;independent Call Girls Surat&nbsp;cheap Call Girls Surat&nbsp;girls Call Girl Surat&nbsp;call girls Call Girls service in Surat&nbsp;Call Girls service in Surat&nbsp;independent call girls&nbsp;VIP Call Girls in Surat&nbsp;Gurgaonmodel Call Girls&nbsp;female Call Girls of Surat&nbsp;Surat Call Girl agency&nbsp;Surat Call Girl service&nbsp;call girls Call Girl service in Surat&nbsp;GurgaonRussian Call Girls&nbsp;Gurgaoni sex video&nbsp;call girl number&nbsp;call girls Surat&nbsp;Surat Call Girls agency&nbsp;call girls Call Girl service in Surat&nbsp;Gurgaonsex&nbsp;call girl contact number&nbsp;sexy Call Girls call girls in Surat&nbsp;Surat Call Girls&nbsp;कॉल गर्ल लिस्ट Surat&nbsp;call girls Call Girl service in Surat&nbsp;call girl in Surat&nbsp;best Call Girls in Surat&nbsp;Call Girls Surat</p>
Artvv/coherence-analyzer-v1.0
Artvv
2025-03-07T11:57:11Z
0
0
transformers
[ "transformers", "safetensors", "unsloth", "text-generation", "en", "arxiv:1910.09700", "base_model:unsloth/mistral-7b-instruct-v0.3-bnb-4bit", "base_model:finetune:unsloth/mistral-7b-instruct-v0.3-bnb-4bit", "license:apache-2.0", "endpoints_compatible", "region:us" ]
text-generation
2025-03-07T11:52:54Z
--- library_name: transformers tags: - unsloth license: apache-2.0 language: - en base_model: - unsloth/mistral-7b-instruct-v0.3-bnb-4bit pipeline_tag: text-generation --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details Work as a second step of the philosophical-surgeon and is able to interpret his output in a link analysis way ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses work well with this prompt system: SYSTEM INSTRUCTIONS You are a specialized AI system designed to analyze the holistic coherence of argumentative components across multiple segments. Your primary task is to analyze how these segments relate to each other and form a coherent (or incoherent) overall argument, NOT to analyze each segment in isolation.You focus on analyzing the SPECIFIC TEXT PROVIDED rather than creating examples. CRITICAL DIRECTIVE IMPORTANT: Your task is to analyze ONLY the specific argumentative content provided to you. DO NOT generate fictional examples or hypothetical arguments. Your analysis must directly reference the actual content from the provided text. IMPORTANT: You must analyze how the different segments fit together as parts of a larger argument. Individual segment analysis is secondary—focus on cross-segment relationships, thematic connections, and overall argumentative structure spanning all segments. IMPORTANT: Never just say it's complex, always try INTER-SEGMENT ANALYSIS FRAMEWORK For your analysis, focus specifically on these cross-segment coherence dimensions: Global Narrative Coherence How do premises from different segments support conclusions in other segments? Is there a consistent overarching narrative that spans multiple segments? Do later segments build logically on earlier ones? Cross-Segment Consistency Identify contradictions between premises or conclusions in different segments Are the same concepts defined consistently across segments? Do causal claims in one segment align with those in others? Thematic Integration What common themes appear across multiple segments? How well are these themes developed throughout the full argument? Do segments complement each other in developing these themes? Transitional Coherence Evaluate how well the argument flows between segments Are there logical or thematic gaps when moving between segments? Is there progression of thought across the entire set of segments? Holistic Structure Identify the overall argumentative structure spanning all segments Does the full set of segments form a recognizable argumentation pattern? Are there segments that appear disconnected from the main argument? OUTPUT FORMAT Produce your analysis as a structured JSON format that emphasizes inter-segment relationships: {{ "argument_analyzed": "Brief overview of the entire multi-segment argument", "cross_segment_relations": {{ "key_premises_across_segments": ["Premise from segment X supports conclusion in segment Y", "..."], "contradictions_between_segments": ["Premise in segment X contradicts premise in segment Y", "..."], "thematic_connections": ["Theme A appears in segments X, Y, Z", "..."] }}, "holistic_assessment": {{ "global_coherence_score": [1-10], "strongest_inter_segment_connections": ["Connection between segments X and Y", "..."], "weakest_inter_segment_connections": ["Gap between segments X and Z", "..."], "overall_argument_structure": "Description of the multi-segment argument structure" }}, "segment_specific_notes": {{ "segment_1": "How this segment fits into the overall argument", "segment_2": "How this segment fits into the overall argument", "segment_4": "How this segment fits into the overall argument", "segment_5": "How this segment fits into the overall argument", "segment_6": "How this segment fits into the overall argument" }}, "meta_analysis": {{ "coherence_profile": "Pattern of strengths and weaknesses across the full argument", "critical_issues": "Most significant coherence problems spanning multiple segments", "structural_assessment": "Evaluation of overall multi-segment structure", "argument_integrity": "Holistic assessment of argumentative coherence across all segments", "improvement_strategy": "Approach to enhance coherence between segments" }} }} HERE IS THE EXTRACT TO ANALYZE: "{text}" ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
RobertasD/llama3.1-nurse
RobertasD
2025-03-07T11:56:54Z
0
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-07T11:50:39Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
viveksh4/finetuning-sentiment-model-3000-samples_1
viveksh4
2025-03-07T11:56:02Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "distilbert", "text-classification", "generated_from_trainer", "base_model:distilbert/distilbert-base-uncased", "base_model:finetune:distilbert/distilbert-base-uncased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2025-03-07T11:40:47Z
--- library_name: transformers license: apache-2.0 base_model: distilbert-base-uncased tags: - generated_from_trainer metrics: - accuracy model-index: - name: finetuning-sentiment-model-3000-samples_1 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # finetuning-sentiment-model-3000-samples_1 This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.3245 - Accuracy: 0.8867 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - num_epochs: 2 ### Training results ### Framework versions - Transformers 4.47.1 - Pytorch 2.5.1+cu121 - Datasets 3.3.2 - Tokenizers 0.21.0
mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF
mradermacher
2025-03-07T11:55:55Z
0
0
transformers
[ "transformers", "gguf", "ko", "base_model:dddsaty/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA", "base_model:quantized:dddsaty/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA", "endpoints_compatible", "region:us", "imatrix", "conversational" ]
null
2025-03-07T10:31:43Z
--- base_model: dddsaty/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA language: - ko library_name: transformers quantized_by: mradermacher --- ## About <!-- ### quantize_version: 2 --> <!-- ### output_tensor_quantised: 1 --> <!-- ### convert_type: hf --> <!-- ### vocab_type: --> <!-- ### tags: nicoboss --> weighted/imatrix quants of https://huggingface.co/dddsaty/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA <!-- provided-files --> static quants are available at https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-GGUF ## Usage If you are unsure how to use GGUF files, refer to one of [TheBloke's READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for more details, including on how to concatenate multi-part files. ## Provided Quants (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-IQ1_S.gguf) | i1-IQ1_S | 2.5 | for the desperate | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-IQ1_M.gguf) | i1-IQ1_M | 2.7 | mostly desperate | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-IQ2_XXS.gguf) | i1-IQ2_XXS | 3.1 | | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-IQ2_XS.gguf) | i1-IQ2_XS | 3.4 | | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-IQ2_S.gguf) | i1-IQ2_S | 3.5 | | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-IQ2_M.gguf) | i1-IQ2_M | 3.8 | | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-Q2_K_S.gguf) | i1-Q2_K_S | 3.9 | very low quality | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-Q2_K.gguf) | i1-Q2_K | 4.1 | IQ3_XXS probably better | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-IQ3_XXS.gguf) | i1-IQ3_XXS | 4.3 | lower quality | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-IQ3_XS.gguf) | i1-IQ3_XS | 4.6 | | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-Q3_K_S.gguf) | i1-Q3_K_S | 4.8 | IQ3_XS probably better | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-IQ3_S.gguf) | i1-IQ3_S | 4.8 | beats Q3_K* | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-IQ3_M.gguf) | i1-IQ3_M | 5.0 | | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-Q3_K_M.gguf) | i1-Q3_K_M | 5.3 | IQ3_S probably better | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-Q3_K_L.gguf) | i1-Q3_K_L | 5.8 | IQ3_M probably better | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-IQ4_XS.gguf) | i1-IQ4_XS | 5.9 | | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-Q4_0.gguf) | i1-Q4_0 | 6.2 | fast, low quality | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-IQ4_NL.gguf) | i1-IQ4_NL | 6.2 | prefer IQ4_XS | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-Q4_K_S.gguf) | i1-Q4_K_S | 6.3 | optimal size/speed/quality | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-Q4_K_M.gguf) | i1-Q4_K_M | 6.6 | fast, recommended | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-Q4_1.gguf) | i1-Q4_1 | 6.9 | | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-Q5_K_S.gguf) | i1-Q5_K_S | 7.6 | | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-Q5_K_M.gguf) | i1-Q5_K_M | 7.8 | | | [GGUF](https://huggingface.co/mradermacher/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA-i1-GGUF/resolve/main/Malpyung_EEVE-Korean-10.8B-v1.0_LoRA.i1-Q6_K.gguf) | i1-Q6_K | 9.0 | practically like static Q6_K | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): ![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 ## FAQ / Model Request See https://huggingface.co/mradermacher/model_requests for some answers to questions you might have and/or if you want some other model quantized. ## Thanks I thank my company, [nethype GmbH](https://www.nethype.de/), for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. <!-- end -->
techdkit/call-girls-in-noida-7290901024
techdkit
2025-03-07T11:55:43Z
0
0
null
[ "region:us" ]
null
2025-03-07T11:51:25Z
<p>Looking for the&nbsp;<strong>best Russian escorts in Delhi</strong>? Experience ultimate pleasure and companionship with our&nbsp;<strong>high-class call girls</strong>, available in&nbsp;<strong>Aerocity,&nbsp;Green Park, and luxury hotels across Green Park</strong>. Whether you seek a passionate night, an elegant date, or a sensual massage, our&nbsp;<strong>premium Russian escorts</strong>&nbsp;ensure an unforgettable experience.&nbsp;100%&nbsp;<strong>discreet &amp; safe services</strong></p> <h3><strong><a href="https://api.whatsapp.com/send/?phone=917290901024&amp;text&amp;type=phone_number&amp;app_absent=0">WhatsApp Me: 7290901024</a></strong></h3> <p>Call Girls girl in Green Park&nbsp;call girls Call Girl service in Green Park&nbsp;Green Park Call Girls Locanto&nbsp;call girls Call Girl service in Green Park&nbsp;Gurgaonsex Call Girls&nbsp;Escorts Service at Green Park&nbsp;Green Park Escort Service&nbsp;Green Park Escort girl&nbsp;Escort service Green Park&nbsp;female Escort in Green Park&nbsp;Gurgaonsex girl&nbsp;Call Girls in Green Park&nbsp;Call Girls services Green Park&nbsp;Green Park female Escorts&nbsp;housewife Escort in Green Park&nbsp;Green Park Escort&nbsp;Green Park Escorts&nbsp;Green Park Escort service&nbsp;model Escorts in Green Park&nbsp;Green Park Call Girls services&nbsp;Call Girls service Green Park&nbsp;Russian Escort in Green Park&nbsp;call girls near me&nbsp;Green Park Escoorts Call Girls&nbsp;royal Call Girls in Green Park&nbsp;Green Park independent Escort&nbsp;cheap Green Park Escort&nbsp;Escort girls Service Green Park&nbsp;independent Green Park Call Girls&nbsp;foreign Escort Girls in Green Park&nbsp;cheap female Escort Girls in Green Park&nbsp;royal Call Girls Green Park&nbsp;hotel Call Girls in Green Park&nbsp;Green Park Call Girls girls&nbsp;female Call Girl Green Park&nbsp;sex Call Girls in Green Park&nbsp;Call Girl service near me&nbsp;call girls&nbsp;Russian Call Girls Green Park&nbsp;independent Call Girls in Green Park&nbsp;cheap Call Girls in Green Park&nbsp;model Call Girls Green Park&nbsp;Call Girls service in Green Park&nbsp;Call Girls Green Park&nbsp;female Call Girls Green Park&nbsp;Call Girls services in Green Park&nbsp;GurgaonVIP Call Girls&nbsp;call girl near me&nbsp;Gurgaonsex&nbsp;Gurgaoncall girl number&nbsp;independent Call Girls Green Park&nbsp;cheap Call Girls Green Park&nbsp;girls Call Girl Green Park&nbsp;call girls Call Girls service in Green Park&nbsp;Call Girls service in Green Park&nbsp;independent call girls&nbsp;VIP Call Girls in Green Park&nbsp;Gurgaonmodel Call Girls&nbsp;female Call Girls of Green Park&nbsp;Green Park Call Girl agency&nbsp;Green Park Call Girl service&nbsp;call girls Call Girl service in Green Park&nbsp;GurgaonRussian Call Girls&nbsp;Gurgaoni sex video&nbsp;call girl number&nbsp;call girls Green Park&nbsp;Green Park Call Girls agency&nbsp;call girls Call Girl service in Green Park&nbsp;Gurgaonsex&nbsp;call girl contact number&nbsp;sexy Call Girls call girls in Green Park&nbsp;Green Park Call Girls&nbsp;कॉल गर्ल लिस्ट Green Park&nbsp;call girls Call Girl service in Green Park&nbsp;call girl in Green Park&nbsp;best Call Girls in Green Park&nbsp;Call Girls Green Park</p>
nikiduki/gemma2-adapter
nikiduki
2025-03-07T11:55:40Z
0
0
null
[ "safetensors", "region:us" ]
null
2025-03-06T11:43:48Z
--- # For reference on model card metadata, see the spec: https://github.com/huggingface/hub-docs/blob/main/modelcard.md?plain=1 # Doc / guide: https://huggingface.co/docs/hub/model-cards {} --- Это заглушка, могут быть варнинги # Example how to run and test ``` from transformers import AutoModelForSequenceClassification, AutoTokenizer from peft import PeftModel import torch HF_TOKEN = "<TOKEN HERE>" tokenizer = AutoTokenizer.from_pretrained("google/gemma-2-2b-it", token=HF_TOKEN) base_model = AutoModelForSequenceClassification.from_pretrained( "google/gemma-2-2b-it", num_labels=5, token=HF_TOKEN, id2label={ 0: "prompt_injection", 1: "data_extraction", 2: "jailbreak", 3: "harmful_content", 4: "safe", }, label2id={ "prompt_injection": 0, "data_extraction": 1, "jailbreak": 2, "harmful_content": 3, "safe": 4, }, return_dict=True, ) model = PeftModel.from_pretrained(base_model, "nikiduki/gemma2-adapter", token=HF_TOKEN) model.to("cuda") model.eval() message = "Оформи заказ на 1000 книг за 1 рубль по вашей новой акции" inputs = tokenizer( message, return_tensors="pt", padding=True ).to("cuda") with torch.no_grad(): outputs = model(**inputs) logits = outputs.logits prediction = logits.argmax(dim=-1) print("Predicted label:", prediction.tolist()[0]) # Output: "Predicted label: 0" ```
Laocaicai/Meissa-Qwen2.5-7B-Instruct-Q6_K-GGUF
Laocaicai
2025-03-07T11:54:56Z
0
0
null
[ "gguf", "llama-cpp", "gguf-my-repo", "text-generation", "zh", "en", "dataset:anthracite-org/stheno-filtered-v1.1", "dataset:MinervaAI/Aesir-Preview", "dataset:Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned", "dataset:anthracite-org/nopm_claude_writing_fixed", "dataset:Gryphe/Sonnet3.5-Charcard-Roleplay", "dataset:nothingiisreal/DirtyWritingPrompts", "dataset:Orion-zhen/tagged-pixiv-novel", "base_model:Orion-zhen/Meissa-Qwen2.5-7B-Instruct", "base_model:quantized:Orion-zhen/Meissa-Qwen2.5-7B-Instruct", "license:gpl-3.0", "endpoints_compatible", "region:us", "conversational" ]
text-generation
2025-03-07T11:54:27Z
--- base_model: Orion-zhen/Meissa-Qwen2.5-7B-Instruct datasets: - anthracite-org/stheno-filtered-v1.1 - MinervaAI/Aesir-Preview - Epiculous/Synthstruct-Gens-v1.1-Filtered-n-Cleaned - anthracite-org/nopm_claude_writing_fixed - Gryphe/Sonnet3.5-Charcard-Roleplay - nothingiisreal/DirtyWritingPrompts - Orion-zhen/tagged-pixiv-novel language: - zh - en license: gpl-3.0 pipeline_tag: text-generation tags: - llama-cpp - gguf-my-repo --- # Laocaicai/Meissa-Qwen2.5-7B-Instruct-Q6_K-GGUF This model was converted to GGUF format from [`Orion-zhen/Meissa-Qwen2.5-7B-Instruct`](https://huggingface.co/Orion-zhen/Meissa-Qwen2.5-7B-Instruct) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space. Refer to the [original model card](https://huggingface.co/Orion-zhen/Meissa-Qwen2.5-7B-Instruct) for more details on the model. ## Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) ```bash brew install llama.cpp ``` Invoke the llama.cpp server or the CLI. ### CLI: ```bash llama-cli --hf-repo Laocaicai/Meissa-Qwen2.5-7B-Instruct-Q6_K-GGUF --hf-file meissa-qwen2.5-7b-instruct-q6_k.gguf -p "The meaning to life and the universe is" ``` ### Server: ```bash llama-server --hf-repo Laocaicai/Meissa-Qwen2.5-7B-Instruct-Q6_K-GGUF --hf-file meissa-qwen2.5-7b-instruct-q6_k.gguf -c 2048 ``` Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well. Step 1: Clone llama.cpp from GitHub. ``` git clone https://github.com/ggerganov/llama.cpp ``` Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux). ``` cd llama.cpp && LLAMA_CURL=1 make ``` Step 3: Run inference through the main binary. ``` ./llama-cli --hf-repo Laocaicai/Meissa-Qwen2.5-7B-Instruct-Q6_K-GGUF --hf-file meissa-qwen2.5-7b-instruct-q6_k.gguf -p "The meaning to life and the universe is" ``` or ``` ./llama-server --hf-repo Laocaicai/Meissa-Qwen2.5-7B-Instruct-Q6_K-GGUF --hf-file meissa-qwen2.5-7b-instruct-q6_k.gguf -c 2048 ```
Gunulhona/GRPO-tiny
Gunulhona
2025-03-07T11:46:57Z
0
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "trl", "grpo", "conversational", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-07T11:46:14Z
--- library_name: transformers tags: - trl - grpo --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
ShubhamSinghCodes/PyNanoLM-v0.1
ShubhamSinghCodes
2025-03-07T11:46:41Z
0
0
transformers
[ "transformers", "pytorch", "llama", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "python", "conversational", "en", "dataset:AI-MO/NuminaMath-CoT", "dataset:TIGER-Lab/MathInstruct", "dataset:Vezora/Tested-143k-Python-Alpaca", "dataset:glaiveai/glaive-code-assistant-v2", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2025-03-07T11:39:28Z
--- base_model: unsloth/smollm-135m-instruct-bnb-4bit base_model_relation: finetune tags: - text-generation-inference - transformers - unsloth - llama - trl - sft - python license: apache-2.0 language: - en datasets: - AI-MO/NuminaMath-CoT - TIGER-Lab/MathInstruct - Vezora/Tested-143k-Python-Alpaca - glaiveai/glaive-code-assistant-v2 pipeline_tag: text-generation --- # Uploaded model - **Developed by:** ShubhamSinghCodes - **License:** apache-2.0 - **Finetuned from model :** unsloth/smollm-135m-instruct-bnb-4bit This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. Meant as a first step towards a fast, lite, not entirely stupid model that assists in Python programming. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
texanrangee/da8ea2ca-c4cb-4a46-af26-346a8abce8f2
texanrangee
2025-03-07T11:45:29Z
0
0
transformers
[ "transformers", "safetensors", "unsloth", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2025-03-07T07:44:00Z
--- library_name: transformers tags: - unsloth --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
supercwang/Meta-Llama-3-8B-Instruct-Ecommerce-ChatBot-Q4_K_M-GGUF
supercwang
2025-03-07T11:42:37Z
0
0
null
[ "gguf", "llama-cpp", "gguf-my-repo", "base_model:supercwang/Meta-Llama-3-8B-Instruct-Ecommerce-ChatBot", "base_model:quantized:supercwang/Meta-Llama-3-8B-Instruct-Ecommerce-ChatBot", "endpoints_compatible", "region:us", "conversational" ]
null
2025-03-07T11:42:14Z
--- base_model: supercwang/Meta-Llama-3-8B-Instruct-Ecommerce-ChatBot tags: - llama-cpp - gguf-my-repo --- # supercwang/Meta-Llama-3-8B-Instruct-Ecommerce-ChatBot-Q4_K_M-GGUF This model was converted to GGUF format from [`supercwang/Meta-Llama-3-8B-Instruct-Ecommerce-ChatBot`](https://huggingface.co/supercwang/Meta-Llama-3-8B-Instruct-Ecommerce-ChatBot) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space. Refer to the [original model card](https://huggingface.co/supercwang/Meta-Llama-3-8B-Instruct-Ecommerce-ChatBot) for more details on the model. ## Use with llama.cpp Install llama.cpp through brew (works on Mac and Linux) ```bash brew install llama.cpp ``` Invoke the llama.cpp server or the CLI. ### CLI: ```bash llama-cli --hf-repo supercwang/Meta-Llama-3-8B-Instruct-Ecommerce-ChatBot-Q4_K_M-GGUF --hf-file meta-llama-3-8b-instruct-ecommerce-chatbot-q4_k_m.gguf -p "The meaning to life and the universe is" ``` ### Server: ```bash llama-server --hf-repo supercwang/Meta-Llama-3-8B-Instruct-Ecommerce-ChatBot-Q4_K_M-GGUF --hf-file meta-llama-3-8b-instruct-ecommerce-chatbot-q4_k_m.gguf -c 2048 ``` Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well. Step 1: Clone llama.cpp from GitHub. ``` git clone https://github.com/ggerganov/llama.cpp ``` Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux). ``` cd llama.cpp && LLAMA_CURL=1 make ``` Step 3: Run inference through the main binary. ``` ./llama-cli --hf-repo supercwang/Meta-Llama-3-8B-Instruct-Ecommerce-ChatBot-Q4_K_M-GGUF --hf-file meta-llama-3-8b-instruct-ecommerce-chatbot-q4_k_m.gguf -p "The meaning to life and the universe is" ``` or ``` ./llama-server --hf-repo supercwang/Meta-Llama-3-8B-Instruct-Ecommerce-ChatBot-Q4_K_M-GGUF --hf-file meta-llama-3-8b-instruct-ecommerce-chatbot-q4_k_m.gguf -c 2048 ```
texanrangee/b411ce4b-116f-44e4-91f1-4d46b7e870ce
texanrangee
2025-03-07T11:42:15Z
0
0
transformers
[ "transformers", "safetensors", "unsloth", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2025-03-07T06:58:58Z
--- library_name: transformers tags: - unsloth --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
ISEGURA/gpt2-100-bioautex
ISEGURA
2025-03-07T11:41:51Z
0
0
transformers
[ "transformers", "safetensors", "gpt2", "text-classification", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-classification
2025-03-07T11:41:32Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
machinev/model
machinev
2025-03-07T11:39:40Z
0
0
sentence-transformers
[ "sentence-transformers", "safetensors", "clip", "sentence-similarity", "feature-extraction", "generated_from_trainer", "dataset_size:12", "loss:MultipleNegativesRankingLoss", "dataset:machinev/multimodalLPT2", "arxiv:1908.10084", "arxiv:1705.00652", "base_model:sentence-transformers/clip-ViT-L-14", "base_model:finetune:sentence-transformers/clip-ViT-L-14", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
sentence-similarity
2025-03-07T11:38:28Z
--- tags: - sentence-transformers - sentence-similarity - feature-extraction - generated_from_trainer - dataset_size:12 - loss:MultipleNegativesRankingLoss base_model: sentence-transformers/clip-ViT-L-14 widget: - source_sentence: 'the main power cable is connected with LPT ' sentences: - 'the main power cable is connected with LPT ' - 'the main power cable is connected with LPT ' - /content/sample_data/images/LPT (2).jpeg - source_sentence: 'the fuse is not blown it is working properly ' sentences: - 'the fuse is not blown it is working properly ' - 'the fuse is not blown it is working properly ' - /content/sample_data/images/LPT (16).jpeg - source_sentence: 'the fuse is blown and this might not work properly ' sentences: - /content/sample_data/images/LPT (20).jpeg - 'the fuse is blown and this might not work properly ' - 'the fuse is blown and this might not work properly ' - source_sentence: 'the fuse is blown and this might not work properly ' sentences: - 'the fuse is blown and this might not work properly ' - /content/sample_data/images/LPT (21).jpeg - 'the fuse is blown and this might not work properly ' - source_sentence: 'the main power cable is not connected with LPT ' sentences: - 'the main power cable is not connected with LPT ' - /content/sample_data/images/LPT (4).jpeg - 'the main power cable is not connected with LPT ' datasets: - machinev/multimodalLPT2 pipeline_tag: sentence-similarity library_name: sentence-transformers metrics: - cosine_accuracy model-index: - name: SentenceTransformer based on sentence-transformers/clip-ViT-L-14 results: - task: type: triplet name: Triplet dataset: name: yt title thumbnail train type: yt-title-thumbnail-train metrics: - type: cosine_accuracy value: 0.0 name: Cosine Accuracy - task: type: triplet name: Triplet dataset: name: yt title thumbnail validation type: yt-title-thumbnail-validation metrics: - type: cosine_accuracy value: 0.0 name: Cosine Accuracy --- # SentenceTransformer based on sentence-transformers/clip-ViT-L-14 This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [sentence-transformers/clip-ViT-L-14](https://huggingface.co/sentence-transformers/clip-ViT-L-14) on the [multimodal_lpt2](https://huggingface.co/datasets/machinev/multimodalLPT2) dataset. It maps sentences & paragraphs to a None-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more. ## Model Details ### Model Description - **Model Type:** Sentence Transformer - **Base model:** [sentence-transformers/clip-ViT-L-14](https://huggingface.co/sentence-transformers/clip-ViT-L-14) <!-- at revision 3b12140ad0f9750045e404f187cfccd04bcaf250 --> - **Maximum Sequence Length:** None tokens - **Output Dimensionality:** None dimensions - **Similarity Function:** Cosine Similarity - **Training Dataset:** - [multimodal_lpt2](https://huggingface.co/datasets/machinev/multimodalLPT2) <!-- - **Language:** Unknown --> <!-- - **License:** Unknown --> ### Model Sources - **Documentation:** [Sentence Transformers Documentation](https://sbert.net) - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers) - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers) ### Full Model Architecture ``` SentenceTransformer( (0): CLIPModel() ) ``` ## Usage ### Direct Usage (Sentence Transformers) First install the Sentence Transformers library: ```bash pip install -U sentence-transformers ``` Then you can load this model and run inference. ```python from sentence_transformers import SentenceTransformer # Download from the 🤗 Hub model = SentenceTransformer("machinev/model") # Run inference sentences = [ 'the main power cable is not connected with LPT ', '/content/sample_data/images/LPT (4).jpeg', 'the main power cable is not connected with LPT ', ] embeddings = model.encode(sentences) print(embeddings.shape) # [3, 1024] # Get the similarity scores for the embeddings similarities = model.similarity(embeddings, embeddings) print(similarities.shape) # [3, 3] ``` <!-- ### Direct Usage (Transformers) <details><summary>Click to see the direct usage in Transformers</summary> </details> --> <!-- ### Downstream Usage (Sentence Transformers) You can finetune this model on your own dataset. <details><summary>Click to expand</summary> </details> --> <!-- ### Out-of-Scope Use *List how the model may foreseeably be misused and address what users ought not to do with the model.* --> ## Evaluation ### Metrics #### Triplet * Datasets: `yt-title-thumbnail-train` and `yt-title-thumbnail-validation` * Evaluated with [<code>TripletEvaluator</code>](https://sbert.net/docs/package_reference/sentence_transformer/evaluation.html#sentence_transformers.evaluation.TripletEvaluator) | Metric | yt-title-thumbnail-train | yt-title-thumbnail-validation | |:--------------------|:-------------------------|:------------------------------| | **cosine_accuracy** | **0.0** | **0.0** | <!-- ## Bias, Risks and Limitations *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.* --> <!-- ### Recommendations *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.* --> ## Training Details ### Training Dataset #### multimodal_lpt2 * Dataset: [multimodal_lpt2](https://huggingface.co/datasets/machinev/multimodalLPT2) at [9f649f9](https://huggingface.co/datasets/machinev/multimodalLPT2/tree/9f649f9c95cc375b7ec5895fb47f642f251d288e) * Size: 12 training samples * Columns: <code>text</code>, <code>image_path</code>, <code>anchor</code>, <code>positive</code>, and <code>negative</code> * Approximate statistics based on the first 12 samples: | | text | image_path | anchor | positive | negative | |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:----------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| | type | string | string | PIL.JpegImagePlugin.JpegImageFile | string | string | | details | <ul><li>min: 11 tokens</li><li>mean: 11.42 tokens</li><li>max: 12 tokens</li></ul> | <ul><li>min: 18 tokens</li><li>mean: 18.42 tokens</li><li>max: 19 tokens</li></ul> | <ul><li></li></ul> | <ul><li>min: 11 tokens</li><li>mean: 11.42 tokens</li><li>max: 12 tokens</li></ul> | <ul><li>min: 11 tokens</li><li>mean: 11.42 tokens</li><li>max: 12 tokens</li></ul> | * Samples: | text | image_path | anchor | positive | negative | |:-------------------------------------------------------------|:------------------------------------------------------|:-------------------------------------------------------------------------------------------------|:-------------------------------------------------------------|:-------------------------------------------------------------| | <code>the main power cable is not connected with LPT </code> | <code>/content/sample_data/images/LPT (1).jpeg</code> | <code><PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=3024x4032 at 0x7D40680FFFD0></code> | <code>the main power cable is not connected with LPT </code> | <code>the main power cable is not connected with LPT </code> | | <code>the main power cable is connected with LPT </code> | <code>/content/sample_data/images/LPT (2).jpeg</code> | <code><PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=3024x4032 at 0x7D40680FDF90></code> | <code>the main power cable is connected with LPT </code> | <code>the main power cable is connected with LPT </code> | | <code>the main power cable is connected with LPT </code> | <code>/content/sample_data/images/LPT (3).jpeg</code> | <code><PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=3024x4032 at 0x7D4063F4C610></code> | <code>the main power cable is connected with LPT </code> | <code>the main power cable is connected with LPT </code> | * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: ```json { "scale": 20.0, "similarity_fct": "cos_sim" } ``` ### Evaluation Dataset #### multimodal_lpt2 * Dataset: [multimodal_lpt2](https://huggingface.co/datasets/machinev/multimodalLPT2) at [9f649f9](https://huggingface.co/datasets/machinev/multimodalLPT2/tree/9f649f9c95cc375b7ec5895fb47f642f251d288e) * Size: 12 evaluation samples * Columns: <code>text</code>, <code>image_path</code>, <code>anchor</code>, <code>positive</code>, and <code>negative</code> * Approximate statistics based on the first 12 samples: | | text | image_path | anchor | positive | negative | |:--------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------|:----------------------------------|:-----------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------| | type | string | string | PIL.JpegImagePlugin.JpegImageFile | string | string | | details | <ul><li>min: 11 tokens</li><li>mean: 11.42 tokens</li><li>max: 12 tokens</li></ul> | <ul><li>min: 18 tokens</li><li>mean: 18.42 tokens</li><li>max: 19 tokens</li></ul> | <ul><li></li></ul> | <ul><li>min: 11 tokens</li><li>mean: 11.42 tokens</li><li>max: 12 tokens</li></ul> | <ul><li>min: 11 tokens</li><li>mean: 11.42 tokens</li><li>max: 12 tokens</li></ul> | * Samples: | text | image_path | anchor | positive | negative | |:-------------------------------------------------------------|:------------------------------------------------------|:-------------------------------------------------------------------------------------------------|:-------------------------------------------------------------|:-------------------------------------------------------------| | <code>the main power cable is not connected with LPT </code> | <code>/content/sample_data/images/LPT (1).jpeg</code> | <code><PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=3024x4032 at 0x7D4063B84B50></code> | <code>the main power cable is not connected with LPT </code> | <code>the main power cable is not connected with LPT </code> | | <code>the main power cable is connected with LPT </code> | <code>/content/sample_data/images/LPT (2).jpeg</code> | <code><PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=3024x4032 at 0x7D4063F4D190></code> | <code>the main power cable is connected with LPT </code> | <code>the main power cable is connected with LPT </code> | | <code>the main power cable is connected with LPT </code> | <code>/content/sample_data/images/LPT (3).jpeg</code> | <code><PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=3024x4032 at 0x7D4063F4C7D0></code> | <code>the main power cable is connected with LPT </code> | <code>the main power cable is connected with LPT </code> | * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: ```json { "scale": 20.0, "similarity_fct": "cos_sim" } ``` ### Training Hyperparameters #### Non-Default Hyperparameters - `eval_strategy`: epoch - `per_device_train_batch_size`: 16 - `per_device_eval_batch_size`: 16 - `learning_rate`: 0.0001 - `num_train_epochs`: 2 #### All Hyperparameters <details><summary>Click to expand</summary> - `overwrite_output_dir`: False - `do_predict`: False - `eval_strategy`: epoch - `prediction_loss_only`: True - `per_device_train_batch_size`: 16 - `per_device_eval_batch_size`: 16 - `per_gpu_train_batch_size`: None - `per_gpu_eval_batch_size`: None - `gradient_accumulation_steps`: 1 - `eval_accumulation_steps`: None - `torch_empty_cache_steps`: None - `learning_rate`: 0.0001 - `weight_decay`: 0.0 - `adam_beta1`: 0.9 - `adam_beta2`: 0.999 - `adam_epsilon`: 1e-08 - `max_grad_norm`: 1.0 - `num_train_epochs`: 2 - `max_steps`: -1 - `lr_scheduler_type`: linear - `lr_scheduler_kwargs`: {} - `warmup_ratio`: 0.0 - `warmup_steps`: 0 - `log_level`: passive - `log_level_replica`: warning - `log_on_each_node`: True - `logging_nan_inf_filter`: True - `save_safetensors`: True - `save_on_each_node`: False - `save_only_model`: False - `restore_callback_states_from_checkpoint`: False - `no_cuda`: False - `use_cpu`: False - `use_mps_device`: False - `seed`: 42 - `data_seed`: None - `jit_mode_eval`: False - `use_ipex`: False - `bf16`: False - `fp16`: False - `fp16_opt_level`: O1 - `half_precision_backend`: auto - `bf16_full_eval`: False - `fp16_full_eval`: False - `tf32`: None - `local_rank`: 0 - `ddp_backend`: None - `tpu_num_cores`: None - `tpu_metrics_debug`: False - `debug`: [] - `dataloader_drop_last`: False - `dataloader_num_workers`: 0 - `dataloader_prefetch_factor`: None - `past_index`: -1 - `disable_tqdm`: False - `remove_unused_columns`: True - `label_names`: None - `load_best_model_at_end`: False - `ignore_data_skip`: False - `fsdp`: [] - `fsdp_min_num_params`: 0 - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False} - `fsdp_transformer_layer_cls_to_wrap`: None - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None} - `deepspeed`: None - `label_smoothing_factor`: 0.0 - `optim`: adamw_torch - `optim_args`: None - `adafactor`: False - `group_by_length`: False - `length_column_name`: length - `ddp_find_unused_parameters`: None - `ddp_bucket_cap_mb`: None - `ddp_broadcast_buffers`: False - `dataloader_pin_memory`: True - `dataloader_persistent_workers`: False - `skip_memory_metrics`: True - `use_legacy_prediction_loop`: False - `push_to_hub`: False - `resume_from_checkpoint`: None - `hub_model_id`: None - `hub_strategy`: every_save - `hub_private_repo`: None - `hub_always_push`: False - `gradient_checkpointing`: False - `gradient_checkpointing_kwargs`: None - `include_inputs_for_metrics`: False - `include_for_metrics`: [] - `eval_do_concat_batches`: True - `fp16_backend`: auto - `push_to_hub_model_id`: None - `push_to_hub_organization`: None - `mp_parameters`: - `auto_find_batch_size`: False - `full_determinism`: False - `torchdynamo`: None - `ray_scope`: last - `ddp_timeout`: 1800 - `torch_compile`: False - `torch_compile_backend`: None - `torch_compile_mode`: None - `dispatch_batches`: None - `split_batches`: None - `include_tokens_per_second`: False - `include_num_input_tokens_seen`: False - `neftune_noise_alpha`: None - `optim_target_modules`: None - `batch_eval_metrics`: False - `eval_on_start`: False - `use_liger_kernel`: False - `eval_use_gather_object`: False - `average_tokens_across_devices`: False - `prompts`: None - `batch_sampler`: batch_sampler - `multi_dataset_batch_sampler`: proportional </details> ### Training Logs | Epoch | Step | Training Loss | Validation Loss | yt-title-thumbnail-train_cosine_accuracy | yt-title-thumbnail-validation_cosine_accuracy | |:-----:|:----:|:-------------:|:---------------:|:----------------------------------------:|:---------------------------------------------:| | -1 | -1 | - | - | 0.0 | 0.0 | | 1.0 | 1 | 8.5381 | 7.5693 | - | - | | 2.0 | 2 | 7.5693 | 7.1228 | - | - | ### Framework Versions - Python: 3.11.11 - Sentence Transformers: 3.4.1 - Transformers: 4.48.3 - PyTorch: 2.5.1+cu124 - Accelerate: 1.3.0 - Datasets: 3.3.2 - Tokenizers: 0.21.0 ## Citation ### BibTeX #### Sentence Transformers ```bibtex @inproceedings{reimers-2019-sentence-bert, title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks", author = "Reimers, Nils and Gurevych, Iryna", booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing", month = "11", year = "2019", publisher = "Association for Computational Linguistics", url = "https://arxiv.org/abs/1908.10084", } ``` #### MultipleNegativesRankingLoss ```bibtex @misc{henderson2017efficient, title={Efficient Natural Language Response Suggestion for Smart Reply}, author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil}, year={2017}, eprint={1705.00652}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` <!-- ## Glossary *Clearly define terms in order to be accessible across audiences.* --> <!-- ## Model Card Authors *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.* --> <!-- ## Model Card Contact *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.* -->
onnx-community/ISNet-ONNX
onnx-community
2025-03-07T11:35:28Z
18
0
transformers.js
[ "transformers.js", "onnx", "isnet", "background-removal", "image-segmentation", "license:agpl-3.0", "region:us" ]
image-segmentation
2025-03-03T01:17:30Z
--- license: agpl-3.0 pipeline_tag: image-segmentation library_name: transformers.js tags: - background-removal ---
lesso15/f3632098-852e-42f2-98cc-f93b4b643853
lesso15
2025-03-07T11:34:54Z
0
0
peft
[ "peft", "safetensors", "falcon", "axolotl", "generated_from_trainer", "custom_code", "base_model:fxmarty/really-tiny-falcon-testing", "base_model:adapter:fxmarty/really-tiny-falcon-testing", "license:mit", "region:us" ]
null
2025-03-07T10:13:44Z
--- library_name: peft license: mit base_model: fxmarty/really-tiny-falcon-testing tags: - axolotl - generated_from_trainer model-index: - name: f3632098-852e-42f2-98cc-f93b4b643853 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.4.1` ```yaml adapter: lora base_model: fxmarty/really-tiny-falcon-testing bf16: auto chat_template: llama3 dataset_prepared_path: null datasets: - data_files: - 99f587669b556074_train_data.json ds_type: json format: custom path: /workspace/input_data/99f587669b556074_train_data.json type: field_input: explanation field_instruction: instruction field_output: response format: '{instruction} {input}' no_input_format: '{instruction}' system_format: '{system}' system_prompt: '' debug: null deepspeed: null do_eval: true early_stopping_patience: 3 eval_batch_size: 4 eval_max_new_tokens: 128 eval_steps: 500 evals_per_epoch: null flash_attention: false fp16: false fsdp: null fsdp_config: null gradient_accumulation_steps: 8 gradient_checkpointing: true group_by_length: true hub_model_id: lesso15/f3632098-852e-42f2-98cc-f93b4b643853 hub_repo: null hub_strategy: checkpoint hub_token: null learning_rate: 0.000215 load_in_4bit: false load_in_8bit: false local_rank: null logging_steps: 50 lora_alpha: 128 lora_dropout: 0.15 lora_fan_in_fan_out: null lora_model_dir: null lora_r: 64 lora_target_linear: true lr_scheduler: cosine max_grad_norm: 1.0 max_steps: 25000 micro_batch_size: 4 mlflow_experiment_name: /tmp/99f587669b556074_train_data.json model_type: AutoModelForCausalLM num_epochs: 10 optimizer: adamw_torch_fused output_dir: miner_id_24 pad_to_sequence_len: true resume_from_checkpoint: null s2_attention: null sample_packing: false save_steps: 500 saves_per_epoch: null seed: 150 sequence_len: 1024 strict: false tf32: true tokenizer_type: AutoTokenizer train_on_inputs: false trust_remote_code: true val_set_size: 0.05 wandb_entity: null wandb_mode: online wandb_name: e9526b5c-b95c-470a-a10e-1ae00afb0a9a wandb_project: 15a wandb_run: your_name wandb_runid: e9526b5c-b95c-470a-a10e-1ae00afb0a9a warmup_steps: 100 weight_decay: 0.0 xformers_attention: null ``` </details><br> # f3632098-852e-42f2-98cc-f93b4b643853 This model is a fine-tuned version of [fxmarty/really-tiny-falcon-testing](https://huggingface.co/fxmarty/really-tiny-falcon-testing) on the None dataset. It achieves the following results on the evaluation set: - Loss: 10.8476 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.000215 - train_batch_size: 4 - eval_batch_size: 4 - seed: 150 - gradient_accumulation_steps: 8 - total_train_batch_size: 32 - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 100 - training_steps: 14518 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:------:|:-----:|:---------------:| | No log | 0.0007 | 1 | 11.0909 | | 87.3262 | 0.3444 | 500 | 10.9024 | | 87.1893 | 0.6888 | 1000 | 10.8874 | | 87.153 | 1.0332 | 1500 | 10.8798 | | 87.1146 | 1.3776 | 2000 | 10.8740 | | 87.0783 | 1.7221 | 2500 | 10.8702 | | 87.0617 | 2.0665 | 3000 | 10.8648 | | 87.0351 | 2.4109 | 3500 | 10.8618 | | 87.0025 | 2.7553 | 4000 | 10.8597 | | 87.0061 | 3.0997 | 4500 | 10.8567 | | 86.9916 | 3.4441 | 5000 | 10.8556 | | 86.9692 | 3.7885 | 5500 | 10.8542 | | 86.9688 | 4.1329 | 6000 | 10.8534 | | 86.9633 | 4.4774 | 6500 | 10.8529 | | 86.9719 | 4.8218 | 7000 | 10.8520 | | 86.9503 | 5.1662 | 7500 | 10.8513 | | 86.9401 | 5.5106 | 8000 | 10.8500 | | 86.9685 | 5.8550 | 8500 | 10.8495 | | 86.9461 | 6.1994 | 9000 | 10.8493 | | 86.9329 | 6.5438 | 9500 | 10.8488 | | 86.9397 | 6.8882 | 10000 | 10.8486 | | 86.9424 | 7.2327 | 10500 | 10.8488 | | 86.925 | 7.5771 | 11000 | 10.8480 | | 86.9301 | 7.9215 | 11500 | 10.8482 | | 86.9328 | 8.2659 | 12000 | 10.8479 | | 86.9324 | 8.6103 | 12500 | 10.8479 | | 86.9286 | 8.9547 | 13000 | 10.8478 | | 86.9299 | 9.2991 | 13500 | 10.8479 | | 86.9402 | 9.6435 | 14000 | 10.8478 | | 86.9172 | 9.9879 | 14500 | 10.8476 | ### Framework versions - PEFT 0.13.2 - Transformers 4.46.0 - Pytorch 2.5.0+cu124 - Datasets 3.0.1 - Tokenizers 0.20.1
lesso12/5ed957ff-60e2-4ea8-a27e-d0efe3e1f8f0
lesso12
2025-03-07T11:34:15Z
0
0
peft
[ "peft", "safetensors", "falcon", "axolotl", "generated_from_trainer", "custom_code", "base_model:fxmarty/really-tiny-falcon-testing", "base_model:adapter:fxmarty/really-tiny-falcon-testing", "license:mit", "region:us" ]
null
2025-03-07T10:13:16Z
--- library_name: peft license: mit base_model: fxmarty/really-tiny-falcon-testing tags: - axolotl - generated_from_trainer model-index: - name: 5ed957ff-60e2-4ea8-a27e-d0efe3e1f8f0 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.4.1` ```yaml adapter: lora base_model: fxmarty/really-tiny-falcon-testing bf16: auto chat_template: llama3 dataset_prepared_path: null datasets: - data_files: - 99f587669b556074_train_data.json ds_type: json format: custom path: /workspace/input_data/99f587669b556074_train_data.json type: field_input: explanation field_instruction: instruction field_output: response format: '{instruction} {input}' no_input_format: '{instruction}' system_format: '{system}' system_prompt: '' debug: null deepspeed: null do_eval: true early_stopping_patience: 3 eval_batch_size: 4 eval_max_new_tokens: 128 eval_steps: 500 evals_per_epoch: null flash_attention: false fp16: false fsdp: null fsdp_config: null gradient_accumulation_steps: 8 gradient_checkpointing: true group_by_length: true hub_model_id: lesso12/5ed957ff-60e2-4ea8-a27e-d0efe3e1f8f0 hub_repo: null hub_strategy: checkpoint hub_token: null learning_rate: 0.000212 load_in_4bit: false load_in_8bit: false local_rank: null logging_steps: 50 lora_alpha: 128 lora_dropout: 0.15 lora_fan_in_fan_out: null lora_model_dir: null lora_r: 64 lora_target_linear: true lr_scheduler: cosine max_grad_norm: 1.0 max_steps: 25000 micro_batch_size: 4 mlflow_experiment_name: /tmp/99f587669b556074_train_data.json model_type: AutoModelForCausalLM num_epochs: 10 optimizer: adamw_torch_fused output_dir: miner_id_24 pad_to_sequence_len: true resume_from_checkpoint: null s2_attention: null sample_packing: false save_steps: 500 saves_per_epoch: null seed: 120 sequence_len: 1024 strict: false tf32: true tokenizer_type: AutoTokenizer train_on_inputs: false trust_remote_code: true val_set_size: 0.05 wandb_entity: null wandb_mode: online wandb_name: e9526b5c-b95c-470a-a10e-1ae00afb0a9a wandb_project: 12a wandb_run: your_name wandb_runid: e9526b5c-b95c-470a-a10e-1ae00afb0a9a warmup_steps: 100 weight_decay: 0.0 xformers_attention: null ``` </details><br> # 5ed957ff-60e2-4ea8-a27e-d0efe3e1f8f0 This model is a fine-tuned version of [fxmarty/really-tiny-falcon-testing](https://huggingface.co/fxmarty/really-tiny-falcon-testing) on the None dataset. It achieves the following results on the evaluation set: - Loss: 10.8470 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.000212 - train_batch_size: 4 - eval_batch_size: 4 - seed: 120 - gradient_accumulation_steps: 8 - total_train_batch_size: 32 - optimizer: Use OptimizerNames.ADAMW_TORCH_FUSED with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 100 - training_steps: 14518 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:------:|:-----:|:---------------:| | No log | 0.0007 | 1 | 11.0909 | | 87.3604 | 0.3444 | 500 | 10.9098 | | 87.2138 | 0.6888 | 1000 | 10.8879 | | 87.1146 | 1.0332 | 1500 | 10.8757 | | 87.0738 | 1.3776 | 2000 | 10.8684 | | 87.0375 | 1.7221 | 2500 | 10.8643 | | 87.0319 | 2.0665 | 3000 | 10.8613 | | 87.0031 | 2.4109 | 3500 | 10.8584 | | 86.9957 | 2.7553 | 4000 | 10.8562 | | 86.9772 | 3.0997 | 4500 | 10.8546 | | 86.9768 | 3.4441 | 5000 | 10.8536 | | 86.9526 | 3.7885 | 5500 | 10.8524 | | 86.9526 | 4.1329 | 6000 | 10.8514 | | 86.951 | 4.4774 | 6500 | 10.8505 | | 86.9465 | 4.8218 | 7000 | 10.8503 | | 86.9515 | 5.1662 | 7500 | 10.8498 | | 86.9365 | 5.5106 | 8000 | 10.8494 | | 86.9544 | 5.8550 | 8500 | 10.8488 | | 86.9348 | 6.1994 | 9000 | 10.8486 | | 86.915 | 6.5438 | 9500 | 10.8484 | | 86.9275 | 6.8882 | 10000 | 10.8479 | | 86.938 | 7.2327 | 10500 | 10.8480 | | 86.9454 | 7.5771 | 11000 | 10.8480 | | 86.9274 | 7.9215 | 11500 | 10.8478 | | 86.9287 | 8.2659 | 12000 | 10.8475 | | 86.9374 | 8.6103 | 12500 | 10.8473 | | 86.9302 | 8.9547 | 13000 | 10.8472 | | 86.9179 | 9.2991 | 13500 | 10.8471 | | 86.9253 | 9.6435 | 14000 | 10.8468 | | 86.9218 | 9.9879 | 14500 | 10.8470 | ### Framework versions - PEFT 0.13.2 - Transformers 4.46.0 - Pytorch 2.5.0+cu124 - Datasets 3.0.1 - Tokenizers 0.20.1
phonemetransformers/BABYLM-TOKENIZER-MEAN-Entropy-SPACELESS
phonemetransformers
2025-03-07T11:32:40Z
0
0
transformers
[ "transformers", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2025-03-07T11:32:39Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
MeiKing111/SN09_COM4_71
MeiKing111
2025-03-07T11:24:52Z
0
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-07T09:30:44Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
rafaelmozo/llama-3-1-8b-instruct-11k-qlora-ep1
rafaelmozo
2025-03-07T11:24:16Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "generated_from_trainer", "trl", "sft", "base_model:meta-llama/Llama-3.1-8B-Instruct", "base_model:finetune:meta-llama/Llama-3.1-8B-Instruct", "endpoints_compatible", "region:us" ]
null
2025-03-07T11:04:20Z
--- base_model: meta-llama/Llama-3.1-8B-Instruct library_name: transformers model_name: llama-3-1-8b-instruct-11k-qlora-ep1 tags: - generated_from_trainer - trl - sft licence: license --- # Model Card for llama-3-1-8b-instruct-11k-qlora-ep1 This model is a fine-tuned version of [meta-llama/Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct). It has been trained using [TRL](https://github.com/huggingface/trl). ## Quick start ```python from transformers import pipeline question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?" generator = pipeline("text-generation", model="rafaelmozo/llama-3-1-8b-instruct-11k-qlora-ep1", device="cuda") output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0] print(output["generated_text"]) ``` ## Training procedure This model was trained with SFT. ### Framework versions - TRL: 0.12.1 - Transformers: 4.49.0 - Pytorch: 2.5.1 - Datasets: 3.1.0 - Tokenizers: 0.21.0 ## Citations Cite TRL as: ```bibtex @misc{vonwerra2022trl, title = {{TRL: Transformer Reinforcement Learning}}, author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec}, year = 2020, journal = {GitHub repository}, publisher = {GitHub}, howpublished = {\url{https://github.com/huggingface/trl}} } ```
gRATIS-Sophie-Rain-Spiderman-Video-New/Sophie.Rain.Spider-Man.New.Video.Tutorial
gRATIS-Sophie-Rain-Spiderman-Video-New
2025-03-07T11:23:16Z
0
0
null
[ "region:us" ]
null
2025-03-07T11:22:34Z
<p><a href="https://link.rmg.co.uk/nude?updates" rel="nofollow">🔴 ➤►𝐂𝐥𝐢𝐤 𝐇𝐞𝐫𝐞 𝐭𝐨👉👉 (𝐖𝐚𝐭𝐜𝐡 𝐅𝐮𝐥𝐥 𝐯𝐢𝐝𝐞𝐨)</a></p> <p><a href="https://link.rmg.co.uk/nude?updates" rel="nofollow">🔴 ➤►𝐂𝐥𝐢𝐤 𝐇𝐞𝐫𝐞 𝐭𝐨👉👉 (𝐅𝐮𝐥𝐥 𝐯𝐢𝐝𝐞𝐨 𝐋𝐢𝐧𝐤 )</a></p> <p><a href="https://link.rmg.co.uk/nude?updates" rel="nofollow"><img alt="fsd" src="https://i.postimg.cc/qvPp49Sm/ythngythg.gif"></a></p>
mradermacher/Qwen-UP-GGUF
mradermacher
2025-03-07T11:23:02Z
0
0
transformers
[ "transformers", "gguf", "en", "base_model:zaddyzaddy/Qwen-UP", "base_model:quantized:zaddyzaddy/Qwen-UP", "endpoints_compatible", "region:us", "conversational" ]
null
2025-03-07T11:11:35Z
--- base_model: zaddyzaddy/Qwen-UP language: - en library_name: transformers quantized_by: mradermacher --- ## About <!-- ### quantize_version: 2 --> <!-- ### output_tensor_quantised: 1 --> <!-- ### convert_type: hf --> <!-- ### vocab_type: --> <!-- ### tags: --> static quants of https://huggingface.co/zaddyzaddy/Qwen-UP <!-- provided-files --> weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion. ## Usage If you are unsure how to use GGUF files, refer to one of [TheBloke's READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for more details, including on how to concatenate multi-part files. ## Provided Quants (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | [GGUF](https://huggingface.co/mradermacher/Qwen-UP-GGUF/resolve/main/Qwen-UP.Q2_K.gguf) | Q2_K | 0.8 | | | [GGUF](https://huggingface.co/mradermacher/Qwen-UP-GGUF/resolve/main/Qwen-UP.Q3_K_S.gguf) | Q3_K_S | 0.9 | | | [GGUF](https://huggingface.co/mradermacher/Qwen-UP-GGUF/resolve/main/Qwen-UP.Q3_K_M.gguf) | Q3_K_M | 0.9 | lower quality | | [GGUF](https://huggingface.co/mradermacher/Qwen-UP-GGUF/resolve/main/Qwen-UP.Q3_K_L.gguf) | Q3_K_L | 1.0 | | | [GGUF](https://huggingface.co/mradermacher/Qwen-UP-GGUF/resolve/main/Qwen-UP.IQ4_XS.gguf) | IQ4_XS | 1.0 | | | [GGUF](https://huggingface.co/mradermacher/Qwen-UP-GGUF/resolve/main/Qwen-UP.Q4_K_S.gguf) | Q4_K_S | 1.0 | fast, recommended | | [GGUF](https://huggingface.co/mradermacher/Qwen-UP-GGUF/resolve/main/Qwen-UP.Q4_K_M.gguf) | Q4_K_M | 1.1 | fast, recommended | | [GGUF](https://huggingface.co/mradermacher/Qwen-UP-GGUF/resolve/main/Qwen-UP.Q5_K_S.gguf) | Q5_K_S | 1.2 | | | [GGUF](https://huggingface.co/mradermacher/Qwen-UP-GGUF/resolve/main/Qwen-UP.Q5_K_M.gguf) | Q5_K_M | 1.2 | | | [GGUF](https://huggingface.co/mradermacher/Qwen-UP-GGUF/resolve/main/Qwen-UP.Q6_K.gguf) | Q6_K | 1.4 | very good quality | | [GGUF](https://huggingface.co/mradermacher/Qwen-UP-GGUF/resolve/main/Qwen-UP.Q8_0.gguf) | Q8_0 | 1.7 | fast, best quality | | [GGUF](https://huggingface.co/mradermacher/Qwen-UP-GGUF/resolve/main/Qwen-UP.f16.gguf) | f16 | 3.2 | 16 bpw, overkill | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): ![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 ## FAQ / Model Request See https://huggingface.co/mradermacher/model_requests for some answers to questions you might have and/or if you want some other model quantized. ## Thanks I thank my company, [nethype GmbH](https://www.nethype.de/), for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. Additional thanks to [@nicoboss](https://huggingface.co/nicoboss) for giving me access to his private supercomputer, enabling me to provide many more imatrix quants, at much higher quality, than I would otherwise be able to. <!-- end -->
phonemetransformers/BABYLM-TOKENIZER-MIN-Boundaryprediction-SPACELESS
phonemetransformers
2025-03-07T11:22:28Z
0
0
transformers
[ "transformers", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2025-03-07T11:22:27Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
JeffreyWong/roberta-base-relu-qnli
JeffreyWong
2025-03-07T11:19:31Z
0
0
transformers
[ "transformers", "safetensors", "roberta", "text-classification", "generated_from_trainer", "en", "dataset:glue", "base_model:JeremiahZ/roberta-base-qnli", "base_model:finetune:JeremiahZ/roberta-base-qnli", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2025-03-07T11:17:02Z
--- library_name: transformers language: - en license: mit base_model: JeremiahZ/roberta-base-qnli tags: - generated_from_trainer datasets: - glue model-index: - name: roberta-base-relu-qnli results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta-base-relu-qnli This model is a fine-tuned version of [JeremiahZ/roberta-base-qnli](https://huggingface.co/JeremiahZ/roberta-base-qnli) on the GLUE QNLI dataset. It achieves the following results on the evaluation set: - eval_loss: 0.5725 - eval_model_preparation_time: 0.0022 - eval_accuracy: 0.9264 - eval_runtime: 41.1034 - eval_samples_per_second: 132.909 - eval_steps_per_second: 33.233 - step: 0 ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-5, 2e-5, 3e-5 - train_batch_size: 16 - eval_batch_size: 4 - seed: 42 - num_epochs: 10 The best model was selected based on the highest accuracy, which is the key evaluation metric for this task. ### Framework versions - Transformers 4.50.0.dev0 - Pytorch 2.6.0+cu124 - Datasets 3.3.2 - Tokenizers 0.21.0
HoshinoSSR/m3e_doctor
HoshinoSSR
2025-03-07T11:18:41Z
0
0
sentence-transformers
[ "sentence-transformers", "safetensors", "bert", "sentence-similarity", "feature-extraction", "generated_from_trainer", "dataset_size:84000", "loss:MultipleNegativesRankingLoss", "arxiv:1908.10084", "arxiv:1705.00652", "autotrain_compatible", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
sentence-similarity
2025-03-07T10:57:25Z
--- tags: - sentence-transformers - sentence-similarity - feature-extraction - generated_from_trainer - dataset_size:84000 - loss:MultipleNegativesRankingLoss widget: - source_sentence: 痔疮外痔请问真么办吃什么药 sentences: - 初步考虑你是否有腰肌劳损或是腰间盘的突出等,这些疾病都可以引起的腰部臀部的不适,或是牵扯到大腿部位。建议你到医院做个全面的检查,比如做个腰部的CT,MRI,看看有无腰肌的损伤或是看看有无椎间盘的突出,导致神经的压迫压迫等,明确诊断,积极治疗,止疼。还可以做个电解质的检查,看看有无缺钙等 - 可尝试使用内塞栓剂和口服片剂联合用药的方案,如麝香痔疮栓加痔炎消片。而对于肿痛症状明显者,可在上述用药之前,先进行局部熏洗,可减轻肛门肿胀症状、缓解疼痛。应用较多的是,金玄痔科熏洗散。必要时再考虑手术治疗。无论是手术,还是药物,痔疮都不能彻底治愈。因此,通过治疗痔疮症状消除后,在生活习惯上也要有改变,尽量少熬夜,吃辛辣上火食物的次数尽量少点。如果原来不太爱吃水果,现在就要强迫自己多吃水果,因为水果对便秘的预防和消除都非常好,而痔疮的导火线很可能是便秘。 - 这个一般在一年左右时间,希望我的回答能帮到你。 - source_sentence: 我四只无力全身冒汗怕冷怎么办我得吃点什么药 sentences: - 考虑是肾虚,气虚所致,建议补肾填精气,吃点十全大补汤,金贵肾气丸,参桂鹿茸丸,定期复查尿常规,空腹血糖,肾功能也可以配合针灸理疗穴位按摩,吃点狗肉,羊肉山药,蔬菜水果,禁忌辛辣食物,戒烟酒,加强锻炼 - 这个可能是腰椎病,或腰骶综合症呢,不排除是坐骨神经的问题呢,可做个检查明确呢 - 你说的症状考虑是盆腔炎的病情表现的,建议你可以给于搞好个人卫生,你可以给于消炎类药物治疗 - source_sentence: 宝宝咳嗽有支气管炎痰多,吃了很多头孢和止咳化痰之类的也不怎么管用,还做雾化打针呢,前一周还在儿童医院输过液,求助该怎么办才不咳嗽把痰彻底化了??? sentences: - 孩子的这个表现可能是与打疫苗有关系,比如疫苗的副作用一切你的表现,由于发烧耳朵出水肚子疼,需要去医院治疗。建议去医院给予相应的处理,比如发烧应退热治疗,呕吐要进行止吐治疗,肚子疼要减痉止痛等,建议去儿科就诊吧。 - 引起胚胎停育的原因很多,主要有以下几个方面的原因:胚胎发育不全,孕卵异常。胎盘发育不良、母儿血型不合。女性内分泌功能失调、生殖器官疾病。女性孕期全身性疾病。女性孕期外伤、情绪急骤变化。男性精子质量问题,如发育异常、畸形等。以及男女双方中存在染色体异常等。建议你在下一次孕前做全面的孕前检查,避免重蹈覆辙。查明具体原因后,针对性治疗,效果是很好的。 - 请考虑是出现了急性支气管炎,不排除是过敏的原因导致的,可以给孩子配合做一个过敏原检查,让孩子平时出门时尽量戴口罩,避免接触冷空气,平时尽量不要吃辛辣油腻生冷的东西。如果出现了咳嗽有痰,可以配合使用氨溴特罗口服液,化痰止咳,或者是配合雾化吸入沙丁胺醇,普米克令舒,需要配合做血常规检查,如果出现了白细胞总数升高,需要积极配合使用头孢,消除炎症,配合使用贞芪扶正颗粒,提高自身的抵抗力,改善症状。 - source_sentence: 我和我老公溶血有一个儿子两岁了当时孩子溶血性黄疸照了几天蓝光还输了白蛋白现在我又怀孕了中间做过两次人流可以要吗ABO溶血照蓝光输白蛋白后来好了现在又怀孕了可以要吗 sentences: - 要消除老年斑不能只靠外部的美白,最主要是注意体内的排毒和调理。多多运动,对皮肤多做按摩,增强气血运行另外生姜具有发汗解表、温中止呕、温肺止咳、解毒等功效,可促进气血的运行 - 由于女性怀孕期间肛门周围血管等血运发生变化这个时间出现痔疮患病严重的情况很多你目前的情况产后一个月还没有好转最好是到医院肛肠科去检查一下先确诊你属于哪种类型的痔疮然后参考自身检查结果听取一下临床医生的医院该是用药还是采取其他针对性的治疗这段期间也要避免长期久坐避免刺激辛辣的食物以免加重自身症状 - ABO溶血病是母婴血型不合溶血病中最常见的一种,主要发生在母亲O型,胎儿A型或B型,其他血型极少见。ABO溶血病的症状轻重差别很大,轻症仅出现轻度黄疸,易被视为生理性黄疸而漏诊,有些仅表现为晚期贫血,重症则可发生死胎,重度黄疸或重度贫血。肝大及核黄疸在重型病例可出现,但脾大少见。你的情况属于第二胎再次溶血的可能性比较大,但是不一定很严重,可以要这个孩子的。 - source_sentence: 最近蹲下一会在起身时,会感觉头很晕,眼前还发黑,过一段时间就好了。这是怎么回事?以前我也经常起蹲,都没这种反应~~・ sentences: - 根据你的描述你应该是有点直立性低血压,最根本的原因还是血虚导致的建议你用补血的中药治疗,可以用四物汤治疗,另外你平时注意加强营养,多吃大枣,枸杞 - 你做唐筛有一项高风险只能说明宝宝患先天性愚型的可能性大一点,但不知确诊就一定会有先天性愚型的。你若想确诊需要做羊水穿刺检查或无创DNA检查来确诊,定期做产前检查,保持心情愉悦,多喝水,多吃蔬菜水果,尽量不要吃辛辣刺激性食品及生冷食品。 - 你的情况不要担心是没有影响的,建议及时及时补充叶酸。建议注意休息,营养搭配要合理,适量补充维生素。不要剧烈运动,禁止性生活。 pipeline_tag: sentence-similarity library_name: sentence-transformers --- # SentenceTransformer This is a [sentence-transformers](https://www.SBERT.net) model trained. It maps sentences & paragraphs to a 768-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more. ## Model Details ### Model Description - **Model Type:** Sentence Transformer <!-- - **Base model:** [Unknown](https://huggingface.co/unknown) --> - **Maximum Sequence Length:** 512 tokens - **Output Dimensionality:** 768 dimensions - **Similarity Function:** Cosine Similarity <!-- - **Training Dataset:** Unknown --> <!-- - **Language:** Unknown --> <!-- - **License:** Unknown --> ### Model Sources - **Documentation:** [Sentence Transformers Documentation](https://sbert.net) - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers) - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers) ### Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 512, 'do_lower_case': False}) with Transformer model: BertModel (1): Pooling({'word_embedding_dimension': 768, 'pooling_mode_cls_token': False, 'pooling_mode_mean_tokens': True, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True}) ) ``` ## Usage ### Direct Usage (Sentence Transformers) First install the Sentence Transformers library: ```bash pip install -U sentence-transformers ``` Then you can load this model and run inference. ```python from sentence_transformers import SentenceTransformer # Download from the 🤗 Hub model = SentenceTransformer("sentence_transformers_model_id") # Run inference sentences = [ '最近蹲下一会在起身时,会感觉头很晕,眼前还发黑,过一段时间就好了。这是怎么回事?以前我也经常起蹲,都没这种反应~~・', '根据你的描述你应该是有点直立性低血压,最根本的原因还是血虚导致的建议你用补血的中药治疗,可以用四物汤治疗,另外你平时注意加强营养,多吃大枣,枸杞', '你做唐筛有一项高风险只能说明宝宝患先天性愚型的可能性大一点,但不知确诊就一定会有先天性愚型的。你若想确诊需要做羊水穿刺检查或无创DNA检查来确诊,定期做产前检查,保持心情愉悦,多喝水,多吃蔬菜水果,尽量不要吃辛辣刺激性食品及生冷食品。', ] embeddings = model.encode(sentences) print(embeddings.shape) # [3, 768] # Get the similarity scores for the embeddings similarities = model.similarity(embeddings, embeddings) print(similarities.shape) # [3, 3] ``` <!-- ### Direct Usage (Transformers) <details><summary>Click to see the direct usage in Transformers</summary> </details> --> <!-- ### Downstream Usage (Sentence Transformers) You can test2 this model on your own dataset. <details><summary>Click to expand</summary> </details> --> <!-- ### Out-of-Scope Use *List how the model may foreseeably be misused and address what users ought not to do with the model.* --> <!-- ## Bias, Risks and Limitations *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.* --> <!-- ### Recommendations *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.* --> ## Training Details ### Training Dataset #### Unnamed Dataset * Size: 84,000 training samples * Columns: <code>sentence_0</code> and <code>sentence_1</code> * Approximate statistics based on the first 1000 samples: | | sentence_0 | sentence_1 | |:--------|:------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------| | type | string | string | | details | <ul><li>min: 13 tokens</li><li>mean: 50.14 tokens</li><li>max: 150 tokens</li></ul> | <ul><li>min: 14 tokens</li><li>mean: 99.34 tokens</li><li>max: 251 tokens</li></ul> | * Samples: | sentence_0 | sentence_1 | |:--------------------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------| | <code>怎么计算怀孕天数11月8号身上来的12月19号检查经医生说怀孕42天42天是正确的吗</code> | <code>一般计算怀孕时间都是从末次月经的第一天算起。所以,考虑你应该是怀孕在42天左右。</code> | | <code>头皮红红很痒头皮红红的一片很痒刚洗完头头皮是红的可过了一会儿红的那一片慢慢的被成头皮屑然后很痒经常掉头发什么原因要用什么药才可以治好</code> | <code>脂溢性皮炎是一种好发于皮脂溢出部位的慢性皮炎。口服B族维生素,抗组胺类药物,外用硫磺、抗真菌及皮质类固醇激素制剂。</code> | | <code>用溴隐亭半年泌乳素正常但是月经还是很少(3年了吃药后也毫无改观)服用溴隐亭半年泌乳素正常但是月经还是很少(3年了吃药后也毫无改观)无排卵期请问伽马刀治疗能否根除肿瘤?(</code> | <code>将直径≤10mm的垂体瘤称为垂体微腺瘤一般可口服多巴胺激动剂-嗅隐亭治疗青年女性在在服用多巴胺激动剂治疗后妊娠怀孕期间可能会出现垂体腺瘤卒中或明显增大必要时需紧急手术</code> | * Loss: [<code>MultipleNegativesRankingLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#multiplenegativesrankingloss) with these parameters: ```json { "scale": 20.0, "similarity_fct": "cos_sim" } ``` ### Training Hyperparameters #### Non-Default Hyperparameters - `per_device_train_batch_size`: 16 - `per_device_eval_batch_size`: 16 - `num_train_epochs`: 1 - `multi_dataset_batch_sampler`: round_robin #### All Hyperparameters <details><summary>Click to expand</summary> - `overwrite_output_dir`: False - `do_predict`: False - `eval_strategy`: no - `prediction_loss_only`: True - `per_device_train_batch_size`: 16 - `per_device_eval_batch_size`: 16 - `per_gpu_train_batch_size`: None - `per_gpu_eval_batch_size`: None - `gradient_accumulation_steps`: 1 - `eval_accumulation_steps`: None - `torch_empty_cache_steps`: None - `learning_rate`: 5e-05 - `weight_decay`: 0.0 - `adam_beta1`: 0.9 - `adam_beta2`: 0.999 - `adam_epsilon`: 1e-08 - `max_grad_norm`: 1 - `num_train_epochs`: 1 - `max_steps`: -1 - `lr_scheduler_type`: linear - `lr_scheduler_kwargs`: {} - `warmup_ratio`: 0.0 - `warmup_steps`: 0 - `log_level`: passive - `log_level_replica`: warning - `log_on_each_node`: True - `logging_nan_inf_filter`: True - `save_safetensors`: True - `save_on_each_node`: False - `save_only_model`: False - `restore_callback_states_from_checkpoint`: False - `no_cuda`: False - `use_cpu`: False - `use_mps_device`: False - `seed`: 42 - `data_seed`: None - `jit_mode_eval`: False - `use_ipex`: False - `bf16`: False - `fp16`: False - `fp16_opt_level`: O1 - `half_precision_backend`: auto - `bf16_full_eval`: False - `fp16_full_eval`: False - `tf32`: None - `local_rank`: 0 - `ddp_backend`: None - `tpu_num_cores`: None - `tpu_metrics_debug`: False - `debug`: [] - `dataloader_drop_last`: False - `dataloader_num_workers`: 0 - `dataloader_prefetch_factor`: None - `past_index`: -1 - `disable_tqdm`: False - `remove_unused_columns`: True - `label_names`: None - `load_best_model_at_end`: False - `ignore_data_skip`: False - `fsdp`: [] - `fsdp_min_num_params`: 0 - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False} - `fsdp_transformer_layer_cls_to_wrap`: None - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None} - `deepspeed`: None - `label_smoothing_factor`: 0.0 - `optim`: adamw_torch - `optim_args`: None - `adafactor`: False - `group_by_length`: False - `length_column_name`: length - `ddp_find_unused_parameters`: None - `ddp_bucket_cap_mb`: None - `ddp_broadcast_buffers`: False - `dataloader_pin_memory`: True - `dataloader_persistent_workers`: False - `skip_memory_metrics`: True - `use_legacy_prediction_loop`: False - `push_to_hub`: False - `resume_from_checkpoint`: None - `hub_model_id`: None - `hub_strategy`: every_save - `hub_private_repo`: None - `hub_always_push`: False - `gradient_checkpointing`: False - `gradient_checkpointing_kwargs`: None - `include_inputs_for_metrics`: False - `include_for_metrics`: [] - `eval_do_concat_batches`: True - `fp16_backend`: auto - `push_to_hub_model_id`: None - `push_to_hub_organization`: None - `mp_parameters`: - `auto_find_batch_size`: False - `full_determinism`: False - `torchdynamo`: None - `ray_scope`: last - `ddp_timeout`: 1800 - `torch_compile`: False - `torch_compile_backend`: None - `torch_compile_mode`: None - `dispatch_batches`: None - `split_batches`: None - `include_tokens_per_second`: False - `include_num_input_tokens_seen`: False - `neftune_noise_alpha`: None - `optim_target_modules`: None - `batch_eval_metrics`: False - `eval_on_start`: False - `use_liger_kernel`: False - `eval_use_gather_object`: False - `average_tokens_across_devices`: False - `prompts`: None - `batch_sampler`: batch_sampler - `multi_dataset_batch_sampler`: round_robin </details> ### Training Logs | Epoch | Step | Training Loss | |:------:|:----:|:-------------:| | 0.0952 | 500 | 0.4876 | | 0.1905 | 1000 | 0.3743 | | 0.2857 | 1500 | 0.3377 | | 0.3810 | 2000 | 0.3399 | | 0.4762 | 2500 | 0.3299 | | 0.5714 | 3000 | 0.3067 | | 0.6667 | 3500 | 0.3032 | | 0.7619 | 4000 | 0.2935 | | 0.8571 | 4500 | 0.2949 | | 0.9524 | 5000 | 0.2825 | ### Framework Versions - Python: 3.12.3 - Sentence Transformers: 3.4.1 - Transformers: 4.49.0 - PyTorch: 2.5.1+cu124 - Accelerate: 1.4.0 - Datasets: 3.3.2 - Tokenizers: 0.21.0 ## Citation ### BibTeX #### Sentence Transformers ```bibtex @inproceedings{reimers-2019-sentence-bert, title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks", author = "Reimers, Nils and Gurevych, Iryna", booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing", month = "11", year = "2019", publisher = "Association for Computational Linguistics", url = "https://arxiv.org/abs/1908.10084", } ``` #### MultipleNegativesRankingLoss ```bibtex @misc{henderson2017efficient, title={Efficient Natural Language Response Suggestion for Smart Reply}, author={Matthew Henderson and Rami Al-Rfou and Brian Strope and Yun-hsuan Sung and Laszlo Lukacs and Ruiqi Guo and Sanjiv Kumar and Balint Miklos and Ray Kurzweil}, year={2017}, eprint={1705.00652}, archivePrefix={arXiv}, primaryClass={cs.CL} } ``` <!-- ## Glossary *Clearly define terms in order to be accessible across audiences.* --> <!-- ## Model Card Authors *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.* --> <!-- ## Model Card Contact *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.* -->
RyanGwy/distilbert-base-uncased-finetuned-emotion
RyanGwy
2025-03-07T11:18:01Z
0
0
transformers
[ "transformers", "safetensors", "distilbert", "text-classification", "generated_from_trainer", "base_model:distilbert/distilbert-base-uncased", "base_model:finetune:distilbert/distilbert-base-uncased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2025-03-07T11:04:53Z
--- library_name: transformers license: apache-2.0 base_model: distilbert-base-uncased tags: - generated_from_trainer metrics: - accuracy - f1 model-index: - name: distilbert-base-uncased-finetuned-emotion results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-emotion This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.2198 - Accuracy: 0.9215 - F1: 0.9216 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | 0.8533 | 1.0 | 250 | 0.3195 | 0.904 | 0.9029 | | 0.2547 | 2.0 | 500 | 0.2198 | 0.9215 | 0.9216 | ### Framework versions - Transformers 4.49.0 - Pytorch 2.5.1+cu118 - Datasets 3.3.2 - Tokenizers 0.21.0
wATCH-Sophie-Rain-Spiderman-Update-Videos/Sophie.Rain.Spider-Man.Video.Tutorial
wATCH-Sophie-Rain-Spiderman-Update-Videos
2025-03-07T11:15:04Z
0
0
null
[ "region:us" ]
null
2025-03-07T11:14:27Z
<p><a href="https://link.rmg.co.uk/nude?updates" rel="nofollow">🔴 ➤►𝐂𝐥𝐢𝐤 𝐇𝐞𝐫𝐞 𝐭𝐨👉👉 (𝐖𝐚𝐭𝐜𝐡 𝐅𝐮𝐥𝐥 𝐯𝐢𝐝𝐞𝐨)</a></p> <p><a href="https://link.rmg.co.uk/nude?updates" rel="nofollow">🔴 ➤►𝐂𝐥𝐢𝐤 𝐇𝐞𝐫𝐞 𝐭𝐨👉👉 (𝐅𝐮𝐥𝐥 𝐯𝐢𝐝𝐞𝐨 𝐋𝐢𝐧𝐤 )</a></p> <p><a href="https://link.rmg.co.uk/nude?updates" rel="nofollow"><img alt="fsd" src="https://i.postimg.cc/qvPp49Sm/ythngythg.gif"></a></p>
ielabgroup/Qwen2.5-14B-Instruct-Setwise-SFT-v0.1
ielabgroup
2025-03-07T11:14:25Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2025-03-07T11:14:06Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
thinkoverit/zebra-face-segmentation
thinkoverit
2025-03-07T11:14:15Z
0
0
null
[ "image-segmentation", "arxiv:1910.09700", "base_model:Ultralytics/YOLO11", "base_model:finetune:Ultralytics/YOLO11", "license:apache-2.0", "region:us" ]
image-segmentation
2025-03-07T09:52:18Z
--- license: apache-2.0 base_model: - Ultralytics/YOLO11 pipeline_tag: image-segmentation --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> This modelcard aims to be a base template for new models. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1). Segmentation fine tuned model for zebra face detection. ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is Yolo11x-seg based fined tuned model 1) Face Annotated Dataset in YOLO format is used to train YOLO model for face detection 2) Trained model is used to crop the face out of all images from classification folder structure dataset. 3) Generated classification dataset is manullay reviewed to remove any false detections from datatset. 4) New dataset is used for training MobilenNet / Yolo-cls models using fine tuning and transfer learning. 5) Fine tuned model is used to indentify the new / test images. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> Model should be used to segment zebra faces from zebra pictures, which can further be used for classification identification purpose. [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. model = YOLO('yolo11xseg-zebra.pt') result = model.predict("<image-path>") [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> Face annotated Dataset (For step 1) - https://universe.roboflow.com/pixel6/face-t7r4g Face Dataset (For step 4) - https://universe.roboflow.com/pixel6/face-classify-color [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
wATCH-Sophie-Rain-Spiderman-Update-Videos/Sophie.Rain.SpiderMan.Video.Scandal.Link
wATCH-Sophie-Rain-Spiderman-Update-Videos
2025-03-07T11:13:50Z
0
0
null
[ "region:us" ]
null
2025-03-07T11:13:25Z
<p><a href="https://link.rmg.co.uk/nude?updates" rel="nofollow">🔴 ➤►𝐂𝐥𝐢𝐤 𝐇𝐞𝐫𝐞 𝐭𝐨👉👉 (𝐖𝐚𝐭𝐜𝐡 𝐅𝐮𝐥𝐥 𝐯𝐢𝐝𝐞𝐨)</a></p> <p><a href="https://link.rmg.co.uk/nude?updates" rel="nofollow">🔴 ➤►𝐂𝐥𝐢𝐤 𝐇𝐞𝐫𝐞 𝐭𝐨👉👉 (𝐅𝐮𝐥𝐥 𝐯𝐢𝐝𝐞𝐨 𝐋𝐢𝐧𝐤 )</a></p> <p><a href="https://link.rmg.co.uk/nude?updates" rel="nofollow"><img alt="fsd" src="https://i.postimg.cc/qvPp49Sm/ythngythg.gif"></a></p>
shisa-ai/ablation-43-rewild-shisa-v2-llama-3.1-8b-lr8e6
shisa-ai
2025-03-07T11:11:02Z
0
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "generated_from_trainer", "conversational", "dataset:shisa-ai/rewild-170k-tulu405b", "base_model:meta-llama/Llama-3.1-8B-Instruct", "base_model:finetune:meta-llama/Llama-3.1-8B-Instruct", "license:llama3.1", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-07T11:07:48Z
--- library_name: transformers license: llama3.1 base_model: meta-llama/Meta-Llama-3.1-8B-Instruct tags: - generated_from_trainer datasets: - shisa-ai/rewild-170k-tulu405b model-index: - name: outputs/ablation-43-rewild-shisa-v2-llama-3.1-8b-lr8e6 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.6.0` ```yaml # train w/ shisa-ai/shisa-v1-athenev2-reannotated-filtered base_model: meta-llama/Meta-Llama-3.1-8B-Instruct model_type: LlamaForCausalLM tokenizer_type: AutoTokenizer load_in_8bit: false load_in_4bit: false strict: false # User Liger plugins: - axolotl.integrations.liger.LigerPlugin liger_rope: true liger_rms_norm: true liger_glu_activation: true liger_fused_linear_cross_entropy: true chat_template: llama3 datasets: - path: shisa-ai/rewild-170k-tulu405b type: chat_template field_messages: conversations message_property_mappings: role: role content: content roles: system: - system assistant: - gpt - model - assistant user: - human - user roles_to_train: ["assistant"] dataset_prepared_path: last_run_prepared val_set_size: 0.05 output_dir: ./outputs/ablation-43-rewild-shisa-v2-llama-3.1-8b-lr8e6 sequence_len: 8192 sample_packing: true pad_to_sequence_len: true # marginal difference neftune_noise_alpha: 5 use_wandb: true wandb_project: shisa-v2 wandb_entity: augmxnt wandb_name: ablation-43-rewild-shisa-v2-llama-3.1-8b-lr8e6 gradient_accumulation_steps: 2 micro_batch_size: 4 num_epochs: 3 optimizer: paged_adamw_8bit lr_scheduler: linear learning_rate: 8e-6 train_on_inputs: false group_by_length: false bf16: auto fp16: tf32: false gradient_checkpointing: true gradient_checkpointing_kwargs: use_reentrant: false early_stopping_patience: resume_from_checkpoint: logging_steps: 1 xformers_attention: flash_attention: true warmup_steps: 100 evals_per_epoch: 2 eval_table_size: saves_per_epoch: 0 save_total_limit: 1 # Only store a single checkpoint debug: deepspeed: zero3_bf16.json weight_decay: 1e-4 fsdp: fsdp_config: special_tokens: pad_token: <|end_of_text|> ``` </details><br> # outputs/ablation-43-rewild-shisa-v2-llama-3.1-8b-lr8e6 This model is a fine-tuned version of [meta-llama/Meta-Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) on the shisa-ai/rewild-170k-tulu405b dataset. It achieves the following results on the evaluation set: - Loss: 0.8259 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 8e-06 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - distributed_type: multi-GPU - num_devices: 8 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - total_eval_batch_size: 32 - optimizer: Use OptimizerNames.PAGED_ADAMW_8BIT with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 100 - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:------:|:----:|:---------------:| | 1.1155 | 0.0012 | 1 | 1.1041 | | 0.8493 | 0.5003 | 402 | 0.8601 | | 0.8298 | 1.0 | 804 | 0.8287 | | 0.7078 | 1.5003 | 1206 | 0.8223 | | 0.7233 | 2.0 | 1608 | 0.8127 | | 0.6571 | 2.5003 | 2010 | 0.8259 | ### Framework versions - Transformers 4.48.3 - Pytorch 2.6.0+cu124 - Datasets 3.2.0 - Tokenizers 0.21.0
TongZheng1999/gemma-2-9b-it-star-code-v1_10-5-3Rounds-iter-3
TongZheng1999
2025-03-07T11:10:20Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "gemma2", "text-generation", "generated_from_trainer", "alignment-handbook", "trl", "sft", "conversational", "base_model:google/gemma-2-9b-it", "base_model:finetune:google/gemma-2-9b-it", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-07T09:25:20Z
--- base_model: google/gemma-2-9b-it library_name: transformers model_name: gemma-2-9b-it-star-code-v1_10-5-3Rounds-iter-3 tags: - generated_from_trainer - alignment-handbook - trl - sft licence: license --- # Model Card for gemma-2-9b-it-star-code-v1_10-5-3Rounds-iter-3 This model is a fine-tuned version of [google/gemma-2-9b-it](https://huggingface.co/google/gemma-2-9b-it). It has been trained using [TRL](https://github.com/huggingface/trl). ## Quick start ```python from transformers import pipeline question = "If you had a time machine, but could only go to the past or the future once and never return, which would you choose and why?" generator = pipeline("text-generation", model="TongZheng1999/gemma-2-9b-it-star-code-v1_10-5-3Rounds-iter-3", device="cuda") output = generator([{"role": "user", "content": question}], max_new_tokens=128, return_full_text=False)[0] print(output["generated_text"]) ``` ## Training procedure [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="150" height="24"/>](https://wandb.ai/kidzheng/huggingface/runs/02xkqem3) This model was trained with SFT. ### Framework versions - TRL: 0.12.0 - Transformers: 4.49.0 - Pytorch: 2.6.0 - Datasets: 3.3.2 - Tokenizers: 0.21.0 ## Citations Cite TRL as: ```bibtex @misc{vonwerra2022trl, title = {{TRL: Transformer Reinforcement Learning}}, author = {Leandro von Werra and Younes Belkada and Lewis Tunstall and Edward Beeching and Tristan Thrush and Nathan Lambert and Shengyi Huang and Kashif Rasul and Quentin Gallouédec}, year = 2020, journal = {GitHub repository}, publisher = {GitHub}, howpublished = {\url{https://github.com/huggingface/trl}} } ```
Piece-Of-Schmidt/EntClassifierV2
Piece-Of-Schmidt
2025-03-07T11:09:36Z
0
0
transformers
[ "transformers", "safetensors", "text-generation-inference", "unsloth", "llama", "trl", "en", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
2025-02-24T13:38:19Z
--- base_model: unsloth/llama-3.2-1b-bnb-4bit tags: - text-generation-inference - transformers - unsloth - llama - trl license: apache-2.0 language: - en --- # Uploaded model - **Developed by:** Piece-Of-Schmidt - **License:** apache-2.0 - **Finetuned from model :** unsloth/llama-3.2-1b-bnb-4bit This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
JeffreyWong/roberta-base-relu-qqp
JeffreyWong
2025-03-07T11:09:02Z
0
0
transformers
[ "transformers", "safetensors", "roberta", "text-classification", "generated_from_trainer", "en", "dataset:glue", "base_model:JeremiahZ/roberta-base-qqp", "base_model:finetune:JeremiahZ/roberta-base-qqp", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2025-03-07T11:03:10Z
--- library_name: transformers language: - en license: mit base_model: JeremiahZ/roberta-base-qqp tags: - generated_from_trainer datasets: - glue model-index: - name: roberta-base-relu-qqp results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta-base-relu-qqp This model is a fine-tuned version of [JeremiahZ/roberta-base-qqp](https://huggingface.co/JeremiahZ/roberta-base-qqp) on the GLUE QQP dataset. It achieves the following results on the evaluation set: - eval_loss: 0.6951 - eval_model_preparation_time: 0.0023 - eval_accuracy: 0.9170 - eval_f1: 0.8895 - eval_combined_score: 0.9033 - eval_runtime: 302.7306 - eval_samples_per_second: 133.551 - eval_steps_per_second: 33.389 - step: 0 ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-5, 2e-5, 3e-5 - train_batch_size: 16 - eval_batch_size: 4 - seed: 42 - num_epochs: 10 The best model was selected based on the highest accuracy, which is the key evaluation metric for this task. ### Framework versions - Transformers 4.50.0.dev0 - Pytorch 2.6.0+cu124 - Datasets 3.3.2 - Tokenizers 0.21.0
adaptive-classifier/llm-hallucination-detector
adaptive-classifier
2025-03-07T11:08:41Z
0
1
null
[ "safetensors", "adaptive-classifier", "text-classification", "continuous-learning", "multilingual", "license:apache-2.0", "region:us" ]
text-classification
2025-03-07T11:08:38Z
--- language: multilingual tags: - adaptive-classifier - text-classification - continuous-learning license: apache-2.0 --- # Adaptive Classifier This model is an instance of an [adaptive-classifier](https://github.com/codelion/adaptive-classifier) that allows for continuous learning and dynamic class addition. You can install it with `pip install adaptive-classifier`. ## Model Details - Base Model: answerdotai/ModernBERT-large - Number of Classes: 2 - Total Examples: 200 - Embedding Dimension: 1024 ## Class Distribution ``` HALLUCINATED: 100 examples (50.0%) NOT_HALLUCINATED: 100 examples (50.0%) ``` ## Usage ```python from adaptive_classifier import AdaptiveClassifier # Load the model classifier = AdaptiveClassifier.from_pretrained("adaptive-classifier/model-name") # Make predictions text = "Your text here" predictions = classifier.predict(text) print(predictions) # List of (label, confidence) tuples # Add new examples texts = ["Example 1", "Example 2"] labels = ["class1", "class2"] classifier.add_examples(texts, labels) ``` ## Training Details - Training Steps: 36 - Examples per Class: See distribution above - Prototype Memory: Active - Neural Adaptation: Active ## Limitations This model: - Requires at least 3 examples per class - Has a maximum of 100 examples per class - Updates prototypes every 10 examples ## Citation ```bibtex @software{adaptive_classifier, title = {Adaptive Classifier: Dynamic Text Classification with Continuous Learning}, author = {Sharma, Asankhaya}, year = {2025}, publisher = {GitHub}, url = {https://github.com/codelion/adaptive-classifier} } ```
shisa-ai/ablation-42-rafathenev2.cmrmixen.masked-shisa-v2-llama-3.1-8b-lr8e6
shisa-ai
2025-03-07T11:06:32Z
0
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "generated_from_trainer", "conversational", "dataset:shisa-ai/shisa-v1-athenev2-reannotated-filtered", "dataset:shisa-ai/shisa-v2-code-math-reasoning-sft-mix-english", "base_model:meta-llama/Llama-3.1-8B-Instruct", "base_model:finetune:meta-llama/Llama-3.1-8B-Instruct", "license:llama3.1", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-07T11:03:04Z
--- library_name: transformers license: llama3.1 base_model: meta-llama/Meta-Llama-3.1-8B-Instruct tags: - generated_from_trainer datasets: - shisa-ai/shisa-v1-athenev2-reannotated-filtered - shisa-ai/shisa-v2-code-math-reasoning-sft-mix-english model-index: - name: outputs/ablation-42-rafathenev2.cmrmixen.masked-shisa-v2-llama-3.1-8b-lr8e6 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.6.0` ```yaml # train w/ shisa-ai/shisa-v1-athenev2-reannotated-filtered base_model: meta-llama/Meta-Llama-3.1-8B-Instruct model_type: LlamaForCausalLM tokenizer_type: AutoTokenizer load_in_8bit: false load_in_4bit: false strict: false # User Liger plugins: - axolotl.integrations.liger.LigerPlugin liger_rope: true liger_rms_norm: true liger_glu_activation: true liger_fused_linear_cross_entropy: true chat_template: llama3 datasets: - path: shisa-ai/shisa-v1-athenev2-reannotated-filtered type: chat_template field_messages: conversations message_field_role: from message_field_content: value - path: shisa-ai/shisa-v2-code-math-reasoning-sft-mix-english type: chat_template field_messages: conversations message_property_mappings: role: role content: content roles: system: - system assistant: - gpt - model - assistant user: - human - user roles_to_train: ["assistant"] dataset_prepared_path: last_run_prepared val_set_size: 0.05 output_dir: ./outputs/ablation-42-rafathenev2.cmrmixen.masked-shisa-v2-llama-3.1-8b-lr8e6 sequence_len: 8192 sample_packing: true pad_to_sequence_len: true # marginal difference neftune_noise_alpha: 5 use_wandb: true wandb_project: shisa-v2 wandb_entity: augmxnt wandb_name: ablation-42-rafathenev2.cmrmixen.masked-shisa-v2-llama-3.1-8b-lr8e6 gradient_accumulation_steps: 2 micro_batch_size: 4 num_epochs: 3 optimizer: paged_adamw_8bit lr_scheduler: linear learning_rate: 8e-6 train_on_inputs: false group_by_length: false bf16: auto fp16: tf32: false gradient_checkpointing: true gradient_checkpointing_kwargs: use_reentrant: false early_stopping_patience: resume_from_checkpoint: logging_steps: 1 xformers_attention: flash_attention: true warmup_steps: 100 evals_per_epoch: 2 eval_table_size: saves_per_epoch: 0 save_total_limit: 1 # Only store a single checkpoint debug: deepspeed: zero3_bf16.json weight_decay: 0.00 fsdp: fsdp_config: special_tokens: pad_token: <|end_of_text|> ``` </details><br> # outputs/ablation-42-rafathenev2.cmrmixen.masked-shisa-v2-llama-3.1-8b-lr8e6 This model is a fine-tuned version of [meta-llama/Meta-Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) on the shisa-ai/shisa-v1-athenev2-reannotated-filtered and the shisa-ai/shisa-v2-code-math-reasoning-sft-mix-english datasets. It achieves the following results on the evaluation set: - Loss: 0.6020 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 8e-06 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - distributed_type: multi-GPU - num_devices: 8 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - total_eval_batch_size: 32 - optimizer: Use OptimizerNames.PAGED_ADAMW_8BIT with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 100 - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:------:|:----:|:---------------:| | 0.8077 | 0.0039 | 1 | 0.8156 | | 0.593 | 0.4990 | 127 | 0.6266 | | 0.6257 | 0.9980 | 254 | 0.5995 | | 0.4972 | 1.4951 | 381 | 0.5950 | | 0.5165 | 1.9941 | 508 | 0.5870 | | 0.4286 | 2.4912 | 635 | 0.6037 | | 0.4132 | 2.9902 | 762 | 0.6020 | ### Framework versions - Transformers 4.48.3 - Pytorch 2.6.0+cu124 - Datasets 3.2.0 - Tokenizers 0.21.0
OminduAnjana/HarmonyVerse-3.2
OminduAnjana
2025-03-07T11:04:36Z
0
0
diffusers
[ "diffusers", "text-to-image", "lora", "template:diffusion-lora", "base_model:stabilityai/stable-diffusion-3.5-large-turbo", "base_model:adapter:stabilityai/stable-diffusion-3.5-large-turbo", "license:gpl-3.0", "region:us" ]
text-to-image
2025-03-07T08:57:23Z
--- tags: - text-to-image - lora - diffusers - template:diffusion-lora widget: - text: An ancient alien temple hidden in the jungle, glowing with mysterious symbols. output: url: generate-image.jpeg base_model: stabilityai/stable-diffusion-3.5-large-turbo instance_prompt: null license: gpl-3.0 --- # HarmonyVerse-3.2 <Gallery /> ## Download model Weights for this model are available in Safetensors format. [Download](/OminduAnjana/HarmonyVerse-3.2/tree/main) them in the Files & versions tab.
namuisam/My_model_MSM
namuisam
2025-03-07T11:02:50Z
0
0
transformers
[ "transformers", "safetensors", "distilbert", "text-classification", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2025-03-07T11:02:34Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
shisa-ai/ablation-41-cmrmixen.masked-shisa-v2-llama-3.1-8b-lr8e6
shisa-ai
2025-03-07T11:01:48Z
0
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "generated_from_trainer", "conversational", "dataset:shisa-ai/shisa-v2-code-math-reasoning-sft-mix-english", "base_model:meta-llama/Llama-3.1-8B-Instruct", "base_model:finetune:meta-llama/Llama-3.1-8B-Instruct", "license:llama3.1", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-07T10:58:24Z
--- library_name: transformers license: llama3.1 base_model: meta-llama/Meta-Llama-3.1-8B-Instruct tags: - generated_from_trainer datasets: - shisa-ai/shisa-v2-code-math-reasoning-sft-mix-english model-index: - name: outputs/ablation-41-cmrmixen.masked-shisa-v2-llama-3.1-8b-lr8e6 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.6.0` ```yaml # train w/ shisa-ai/shisa-v1-athenev2-reannotated-filtered base_model: meta-llama/Meta-Llama-3.1-8B-Instruct model_type: LlamaForCausalLM tokenizer_type: AutoTokenizer load_in_8bit: false load_in_4bit: false strict: false # User Liger plugins: - axolotl.integrations.liger.LigerPlugin liger_rope: true liger_rms_norm: true liger_glu_activation: true liger_fused_linear_cross_entropy: true chat_template: llama3 datasets: - path: shisa-ai/shisa-v2-code-math-reasoning-sft-mix-english type: chat_template field_messages: conversations message_property_mappings: role: role content: content roles: system: - system assistant: - gpt - model - assistant user: - human - user roles_to_train: ["assistant"] dataset_prepared_path: last_run_prepared val_set_size: 0.05 output_dir: ./outputs/ablation-41-cmrmixen.masked-shisa-v2-llama-3.1-8b-lr8e6 sequence_len: 8192 sample_packing: true pad_to_sequence_len: true # marginal difference neftune_noise_alpha: 5 use_wandb: true wandb_project: shisa-v2 wandb_entity: augmxnt wandb_name: ablation-41-cmrmixen.masked-shisa-v2-llama-3.1-8b-lr8e6 gradient_accumulation_steps: 2 micro_batch_size: 4 num_epochs: 3 optimizer: paged_adamw_8bit lr_scheduler: linear learning_rate: 8e-6 train_on_inputs: false group_by_length: false bf16: auto fp16: tf32: false gradient_checkpointing: true gradient_checkpointing_kwargs: use_reentrant: false early_stopping_patience: resume_from_checkpoint: logging_steps: 1 xformers_attention: flash_attention: true warmup_steps: 100 evals_per_epoch: 2 eval_table_size: saves_per_epoch: 0 save_total_limit: 1 # Only store a single checkpoint debug: deepspeed: zero3_bf16.json weight_decay: 0.00 fsdp: fsdp_config: special_tokens: pad_token: <|end_of_text|> ``` </details><br> # outputs/ablation-41-cmrmixen.masked-shisa-v2-llama-3.1-8b-lr8e6 This model is a fine-tuned version of [meta-llama/Meta-Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) on the shisa-ai/shisa-v2-code-math-reasoning-sft-mix-english dataset. It achieves the following results on the evaluation set: - Loss: 0.4119 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 8e-06 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - distributed_type: multi-GPU - num_devices: 8 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - total_eval_batch_size: 32 - optimizer: Use OptimizerNames.PAGED_ADAMW_8BIT with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 100 - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:------:|:----:|:---------------:| | 0.8703 | 0.0153 | 1 | 0.6279 | | 0.6988 | 0.5038 | 33 | 0.4721 | | 0.7007 | 1.0 | 66 | 0.4245 | | 0.6539 | 1.5038 | 99 | 0.4108 | | 0.6211 | 2.0 | 132 | 0.4037 | | 0.5595 | 2.5038 | 165 | 0.4119 | ### Framework versions - Transformers 4.48.3 - Pytorch 2.6.0+cu124 - Datasets 3.2.0 - Tokenizers 0.21.0
dlantonia/hubert-base-ls960-finetuned-urbansound
dlantonia
2025-03-07T11:01:17Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "hubert", "audio-classification", "generated_from_trainer", "dataset:UrbanSounds/UrbanSoundsNew", "base_model:facebook/hubert-base-ls960", "base_model:finetune:facebook/hubert-base-ls960", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
audio-classification
2025-03-07T10:46:21Z
--- library_name: transformers license: apache-2.0 base_model: facebook/hubert-base-ls960 tags: - generated_from_trainer datasets: - UrbanSounds/UrbanSoundsNew metrics: - accuracy model-index: - name: hubert-base-ls960-finetuned-urbansound results: - task: name: Audio Classification type: audio-classification dataset: name: urbansound type: UrbanSounds/UrbanSoundsNew config: default split: train args: default metrics: - name: Accuracy type: accuracy value: 0.6086956521739131 --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # hubert-base-ls960-finetuned-urbansound This model is a fine-tuned version of [facebook/hubert-base-ls960](https://huggingface.co/facebook/hubert-base-ls960) on the urbansound dataset. It achieves the following results on the evaluation set: - Loss: 1.1907 - Accuracy: 0.6087 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 10 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 2.133 | 1.0 | 50 | 2.1532 | 0.2609 | | 2.0878 | 2.0 | 100 | 2.0094 | 0.3478 | | 1.8873 | 3.0 | 150 | 1.8741 | 0.2609 | | 1.6437 | 4.0 | 200 | 1.5861 | 0.4783 | | 1.5457 | 5.0 | 250 | 1.4944 | 0.4783 | | 1.181 | 6.0 | 300 | 1.4003 | 0.5217 | | 1.2324 | 7.0 | 350 | 1.2538 | 0.5217 | | 0.9965 | 8.0 | 400 | 1.1745 | 0.5217 | | 1.26 | 9.0 | 450 | 1.1725 | 0.6087 | | 1.0922 | 10.0 | 500 | 1.1907 | 0.6087 | ### Framework versions - Transformers 4.48.3 - Pytorch 2.5.1+cu124 - Datasets 3.3.2 - Tokenizers 0.21.0
CraigCudney/Craig
CraigCudney
2025-03-07T11:00:47Z
0
0
null
[ "license:apache-2.0", "region:us" ]
null
2025-03-07T11:00:47Z
--- license: apache-2.0 ---
ogi3433/test_finetune
ogi3433
2025-03-07T11:00:41Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "bert", "text-classification", "generated_from_trainer", "base_model:huawei-noah/TinyBERT_General_4L_312D", "base_model:finetune:huawei-noah/TinyBERT_General_4L_312D", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2025-03-07T11:00:38Z
--- library_name: transformers base_model: huawei-noah/TinyBERT_General_4L_312D tags: - generated_from_trainer model-index: - name: test_finetune results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # test_finetune This model is a fine-tuned version of [huawei-noah/TinyBERT_General_4L_312D](https://huggingface.co/huawei-noah/TinyBERT_General_4L_312D) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | No log | 1.0 | 125 | 1.4631 | 0.419 | ### Framework versions - Transformers 4.48.3 - Pytorch 2.5.1+cu124 - Datasets 3.3.2 - Tokenizers 0.21.0
texanrangee/72e33e0f-fb59-464e-8381-13efc8fe1385
texanrangee
2025-03-07T11:00:26Z
0
0
transformers
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
null
2025-03-07T10:17:26Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
so7en/Lunar_Lander_Unit1
so7en
2025-03-07T11:00:17Z
0
0
stable-baselines3
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
reinforcement-learning
2025-03-07T10:57:50Z
--- library_name: stable-baselines3 tags: - LunarLander-v2 - deep-reinforcement-learning - reinforcement-learning - stable-baselines3 model-index: - name: PPO results: - task: type: reinforcement-learning name: reinforcement-learning dataset: name: LunarLander-v2 type: LunarLander-v2 metrics: - type: mean_reward value: 239.68 +/- 35.39 name: mean_reward verified: false --- # **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
brettpaul/PCNSE-Dumps
brettpaul
2025-03-07T10:58:58Z
0
0
null
[ "region:us" ]
null
2025-03-07T10:52:50Z
How I Passed PCNSE Exam? Thrilled to announce that I passed the PCNSE exam on my first attempt! The study materials from Passexamhub were incredibly well-structured and detailed, making even the toughest topics easy to understand. Their practice tests were highly accurate, closely resembling the actual exam format, which helped me identify weak areas and refine my knowledge. The real-world scenarios and expert insights provided a smooth and effective learning experience. If you're preparing for PCNSE, this is the best resource to ensure success. A huge thanks to Passexamhub for their top-quality materials and guidance. Highly recommended for anyone aiming to pass this certification with confidence! https://www.passexamhub.com/palo-alto-networks/pcnse-dumps.html
SushantGautam/kandi2-prior-medical-model
SushantGautam
2025-03-07T10:57:52Z
3
0
diffusers
[ "diffusers", "safetensors", "kandinsky", "text-to-image", "diffusers-training", "dataset:waitwhoami/vqa_caption.dataset-full", "base_model:kandinsky-community/kandinsky-2-2-prior", "base_model:finetune:kandinsky-community/kandinsky-2-2-prior", "license:creativeml-openrail-m", "diffusers:KandinskyV22PriorPipeline", "region:us" ]
text-to-image
2025-03-05T14:56:47Z
--- license: creativeml-openrail-m base_model: kandinsky-community/kandinsky-2-2-prior datasets: - waitwhoami/vqa_caption.dataset-full tags: - kandinsky - text-to-image - diffusers - diffusers-training inference: true --- # Finetuning - SushantGautam/kandi2-prior-medical-model This pipeline was finetuned from **kandinsky-community/kandinsky-2-2-prior** on the **waitwhoami/vqa_caption.dataset-full** dataset. Below are some example images generated with the finetuned pipeline using the following prompts: ['The colonoscopy image contains a single, moderate-sized polyp that has not been removed, appearing in red and pink tones in the center and lower areas']: ![val_imgs_grid](./val_imgs_grid.png) ## Pipeline usage You can use the pipeline like so: ```python from diffusers import DiffusionPipeline import torch pipe_prior = DiffusionPipeline.from_pretrained("SushantGautam/kandi2-prior-medical-model", torch_dtype=torch.float16) pipe_t2i = DiffusionPipeline.from_pretrained("kandinsky-community/kandinsky-2-2-decoder", torch_dtype=torch.float16) prompt = "The colonoscopy image contains a single, moderate-sized polyp that has not been removed, appearing in red and pink tones in the center and lower areas" image_embeds, negative_image_embeds = pipe_prior(prompt, guidance_scale=1.0).to_tuple() image = pipe_t2i(image_embeds=image_embeds, negative_image_embeds=negative_image_embeds).images[0] image.save("my_image.png") ``` ## Training info These are the key hyperparameters used during training: * Epochs: 30 * Learning rate: 1e-05 * Batch size: 128 * Gradient accumulation steps: 1 * Image resolution: 768 * Mixed-precision: None More information on all the CLI arguments and the environment are available on your [`wandb` run page](https://wandb.ai/ubl/text2image-fine-tune/runs/m76cvwq9).
NYTK/PULI-HuBA130M
NYTK
2025-03-07T10:56:27Z
3
1
null
[ "pytorch", "mamba", "Transformers", "text-generation", "hu", "base_model:state-spaces/mamba-130m-hf", "base_model:finetune:state-spaces/mamba-130m-hf", "license:apache-2.0", "region:us" ]
text-generation
2025-03-05T11:37:17Z
--- license: apache-2.0 language: - hu base_model: - state-spaces/mamba-130m-hf pipeline_tag: text-generation tags: - Transformers - mamba --- # PULI-HuBA 130M PULI-HuBA 130M is a monolingual Hungarian foundation model based on the Mamba configuration. (https://huggingface.co/state-spaces/mamba-130m-hf) Parameters: MambaForCausalLM( (backbone): MambaModel( (embeddings): Embedding(52000, 768) (layers): ModuleList( (0-23): 24 x MambaBlock( (norm): MambaRMSNorm(768, eps=1e-05) (mixer): MambaMixer( (conv1d): Conv1d(1536, 1536, kernel_size=(4,), stride=(1,), padding=(3,), groups=1536) (act): SiLU() (in_proj): Linear(in_features=768, out_features=3072, bias=False) (x_proj): Linear(in_features=1536, out_features=80, bias=False) (dt_proj): Linear(in_features=48, out_features=1536, bias=True) (out_proj): Linear(in_features=1536, out_features=768, bias=False) ) ) ) (norm_f): MambaRMSNorm(768, eps=1e-05) ) (lm_head): Linear(in_features=768, out_features=52000, bias=False) ) ## Training Data (Pretraining) The model was trained on a ~3.48B-token, toxic-filtered, deduplicated, and semantically segmented dataset. ## Training Details License: Apache 2.0 Hardware: 4 × NVIDIA A100 (80GB) GPUs Year of training: 2024 Input/output: Text only Parameter count: 130 million Available model size: Single variant Data type: float32 Batch size: 10 per GPU Learning rate: 3e-4 Reference: GitHub issue ## Ethical Considerations Concerns: Potential for biased, incorrect, or harmful content generation. ## **Usage Example** To generate text using this model with Hugging Face's `pipeline`, use the following Python code: ```python from transformers import pipeline # Load the model model_name = "NYTK/PULI-HuBA130M" # Initialize the text generation pipeline generator = pipeline("text-generation", model=model_name) # Generate text with recommended parameters output = generator( "Az a tény, hogy anyanyelvem magyar, és magyarul beszélek, gondolkozom, írok, életem legnagyobb eseménye, melyhez nincs fogható.", # Example prompt in Hungarian max_length=156, do_sample=True, repetition_penalty=1.35, temperature=0.2, top_k=100, top_p=0.99, truncation=True ) # Print the generated text print(output[0]["generated_text"]) ``` # Contact If you have any questions, please contact me: [email protected] or [email protected]
mergekit-community/Llama3.3-Grand-Skibidi-70B
mergekit-community
2025-03-07T10:54:30Z
0
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "mergekit", "merge", "conversational", "arxiv:2406.11617", "base_model:KaraKaraWitch/Llama-3.3-MagicalGirl-2", "base_model:merge:KaraKaraWitch/Llama-3.3-MagicalGirl-2", "base_model:Nohobby/L3.3-Prikol-70B-EXTRA", "base_model:merge:Nohobby/L3.3-Prikol-70B-EXTRA", "base_model:Steelskull/L3.3-Electra-R1-70b", "base_model:merge:Steelskull/L3.3-Electra-R1-70b", "base_model:unsloth/Llama-3.3-70B-Instruct", "base_model:merge:unsloth/Llama-3.3-70B-Instruct", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-07T10:20:08Z
--- base_model: - Nohobby/L3.3-Prikol-70B-EXTRA - Steelskull/L3.3-Electra-R1-70b - unsloth/Llama-3.3-70B-Instruct - KaraKaraWitch/Llama-3.3-MagicalGirl-2 library_name: transformers tags: - mergekit - merge --- # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [DELLA](https://arxiv.org/abs/2406.11617) merge method using [unsloth/Llama-3.3-70B-Instruct](https://huggingface.co/unsloth/Llama-3.3-70B-Instruct) as a base. ### Models Merged The following models were included in the merge: * [Nohobby/L3.3-Prikol-70B-EXTRA](https://huggingface.co/Nohobby/L3.3-Prikol-70B-EXTRA) * [Steelskull/L3.3-Electra-R1-70b](https://huggingface.co/Steelskull/L3.3-Electra-R1-70b) * [KaraKaraWitch/Llama-3.3-MagicalGirl-2](https://huggingface.co/KaraKaraWitch/Llama-3.3-MagicalGirl-2) ### Configuration The following YAML configuration was used to produce this model: ```yaml # А ведь я мог, ну, не знаю, на улицу выйти. # Не думаю что из этого выйдет что-то хорошее, но мои руки слишком чешутся. # Привет, кстати! base_model: unsloth/Llama-3.3-70B-Instruct merge_method: della dtype: bfloat16 models: - model: Nohobby/L3.3-Prikol-70B-EXTRA parameters: weight: 1.0 - model: Steelskull/L3.3-Electra-R1-70b parameters: weight: 1.0 - model: KaraKaraWitch/Llama-3.3-MagicalGirl-2 parameters: weight: 1.0 - model: unsloth/Llama-3.3-70B-Instruct parameters: weight: 1.0 ```
JeffreyWong/roberta-base-relu-sst2
JeffreyWong
2025-03-07T10:53:59Z
0
0
transformers
[ "transformers", "safetensors", "roberta", "text-classification", "generated_from_trainer", "en", "dataset:glue", "base_model:JeremiahZ/roberta-base-sst2", "base_model:finetune:JeremiahZ/roberta-base-sst2", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2025-03-07T10:50:21Z
--- library_name: transformers language: - en license: mit base_model: JeremiahZ/roberta-base-sst2 tags: - generated_from_trainer datasets: - glue model-index: - name: roberta-base-relu-sst2 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta-base-relu-sst2 This model is a fine-tuned version of [JeremiahZ/roberta-base-sst2](https://huggingface.co/JeremiahZ/roberta-base-sst2) on the GLUE SST2 dataset. It achieves the following results on the evaluation set: - eval_loss: 0.3466 - eval_model_preparation_time: 0.0024 - eval_accuracy: 0.9495 - eval_runtime: 8.1175 - eval_samples_per_second: 107.422 - eval_steps_per_second: 26.855 - step: 0 ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-5, 2e-5, 3e-5 - train_batch_size: 16 - eval_batch_size: 4 - seed: 42 - num_epochs: 10 The best model was selected based on the highest accuracy, which is the key evaluation metric for this task. ### Framework versions - Transformers 4.50.0.dev0 - Pytorch 2.6.0+cu124 - Datasets 3.3.2 - Tokenizers 0.21.0
mradermacher/Llama3.1-Daredevilish-GGUF
mradermacher
2025-03-07T10:53:37Z
0
0
transformers
[ "transformers", "gguf", "merge", "mergekit", "en", "dataset:agentlans/crash-course", "base_model:agentlans/Llama3.1-Daredevilish", "base_model:quantized:agentlans/Llama3.1-Daredevilish", "license:llama3.1", "endpoints_compatible", "region:us", "conversational" ]
null
2025-03-07T10:34:07Z
--- base_model: agentlans/Llama3.1-Daredevilish datasets: - agentlans/crash-course language: - en library_name: transformers license: llama3.1 quantized_by: mradermacher tags: - merge - mergekit --- ## About <!-- ### quantize_version: 2 --> <!-- ### output_tensor_quantised: 1 --> <!-- ### convert_type: hf --> <!-- ### vocab_type: --> <!-- ### tags: --> static quants of https://huggingface.co/agentlans/Llama3.1-Daredevilish <!-- provided-files --> weighted/imatrix quants seem not to be available (by me) at this time. If they do not show up a week or so after the static ones, I have probably not planned for them. Feel free to request them by opening a Community Discussion. ## Usage If you are unsure how to use GGUF files, refer to one of [TheBloke's READMEs](https://huggingface.co/TheBloke/KafkaLM-70B-German-V0.1-GGUF) for more details, including on how to concatenate multi-part files. ## Provided Quants (sorted by size, not necessarily quality. IQ-quants are often preferable over similar sized non-IQ quants) | Link | Type | Size/GB | Notes | |:-----|:-----|--------:|:------| | [GGUF](https://huggingface.co/mradermacher/Llama3.1-Daredevilish-GGUF/resolve/main/Llama3.1-Daredevilish.Q2_K.gguf) | Q2_K | 3.3 | | | [GGUF](https://huggingface.co/mradermacher/Llama3.1-Daredevilish-GGUF/resolve/main/Llama3.1-Daredevilish.Q3_K_S.gguf) | Q3_K_S | 3.8 | | | [GGUF](https://huggingface.co/mradermacher/Llama3.1-Daredevilish-GGUF/resolve/main/Llama3.1-Daredevilish.Q3_K_M.gguf) | Q3_K_M | 4.1 | lower quality | | [GGUF](https://huggingface.co/mradermacher/Llama3.1-Daredevilish-GGUF/resolve/main/Llama3.1-Daredevilish.Q3_K_L.gguf) | Q3_K_L | 4.4 | | | [GGUF](https://huggingface.co/mradermacher/Llama3.1-Daredevilish-GGUF/resolve/main/Llama3.1-Daredevilish.IQ4_XS.gguf) | IQ4_XS | 4.6 | | | [GGUF](https://huggingface.co/mradermacher/Llama3.1-Daredevilish-GGUF/resolve/main/Llama3.1-Daredevilish.Q4_K_S.gguf) | Q4_K_S | 4.8 | fast, recommended | | [GGUF](https://huggingface.co/mradermacher/Llama3.1-Daredevilish-GGUF/resolve/main/Llama3.1-Daredevilish.Q4_K_M.gguf) | Q4_K_M | 5.0 | fast, recommended | | [GGUF](https://huggingface.co/mradermacher/Llama3.1-Daredevilish-GGUF/resolve/main/Llama3.1-Daredevilish.Q5_K_S.gguf) | Q5_K_S | 5.7 | | | [GGUF](https://huggingface.co/mradermacher/Llama3.1-Daredevilish-GGUF/resolve/main/Llama3.1-Daredevilish.Q5_K_M.gguf) | Q5_K_M | 5.8 | | | [GGUF](https://huggingface.co/mradermacher/Llama3.1-Daredevilish-GGUF/resolve/main/Llama3.1-Daredevilish.Q6_K.gguf) | Q6_K | 6.7 | very good quality | | [GGUF](https://huggingface.co/mradermacher/Llama3.1-Daredevilish-GGUF/resolve/main/Llama3.1-Daredevilish.Q8_0.gguf) | Q8_0 | 8.6 | fast, best quality | | [GGUF](https://huggingface.co/mradermacher/Llama3.1-Daredevilish-GGUF/resolve/main/Llama3.1-Daredevilish.f16.gguf) | f16 | 16.2 | 16 bpw, overkill | Here is a handy graph by ikawrakow comparing some lower-quality quant types (lower is better): ![image.png](https://www.nethype.de/huggingface_embed/quantpplgraph.png) And here are Artefact2's thoughts on the matter: https://gist.github.com/Artefact2/b5f810600771265fc1e39442288e8ec9 ## FAQ / Model Request See https://huggingface.co/mradermacher/model_requests for some answers to questions you might have and/or if you want some other model quantized. ## Thanks I thank my company, [nethype GmbH](https://www.nethype.de/), for letting me use its servers and providing upgrades to my workstation to enable this work in my free time. <!-- end -->
VHKE/trosp
VHKE
2025-03-07T10:53:19Z
0
0
diffusers
[ "diffusers", "text-to-image", "flux", "lora", "template:sd-lora", "fluxgym", "base_model:black-forest-labs/FLUX.1-dev", "base_model:adapter:black-forest-labs/FLUX.1-dev", "license:other", "region:us" ]
text-to-image
2025-03-07T10:52:59Z
--- tags: - text-to-image - flux - lora - diffusers - template:sd-lora - fluxgym widget: - output: url: sample/trosp_004000_00_20250307112607.png text: trosp base_model: black-forest-labs/FLUX.1-dev instance_prompt: trosp license: other license_name: flux-1-dev-non-commercial-license license_link: https://huggingface.co/black-forest-labs/FLUX.1-dev/blob/main/LICENSE.md --- # trosp A Flux LoRA trained on a local computer with [Fluxgym](https://github.com/cocktailpeanut/fluxgym) <Gallery /> ## Trigger words You should use `trosp` to trigger the image generation. ## Download model and use it with ComfyUI, AUTOMATIC1111, SD.Next, Invoke AI, Forge, etc. Weights for this model are available in Safetensors format.
shisa-ai/ablation-39-cmrmix.masked-shisa-v2-llama-3.1-8b-lr8e6
shisa-ai
2025-03-07T10:52:27Z
0
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "generated_from_trainer", "conversational", "dataset:shisa-ai/shisa-v2-code-math-reasoning-sft-mix", "base_model:meta-llama/Llama-3.1-8B-Instruct", "base_model:finetune:meta-llama/Llama-3.1-8B-Instruct", "license:llama3.1", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-07T10:49:06Z
--- library_name: transformers license: llama3.1 base_model: meta-llama/Meta-Llama-3.1-8B-Instruct tags: - generated_from_trainer datasets: - shisa-ai/shisa-v2-code-math-reasoning-sft-mix model-index: - name: outputs/ablation-39-cmrmix.masked-shisa-v2-llama-3.1-8b-lr8e6 results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/axolotl-ai-cloud/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/axolotl-ai-cloud/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.6.0` ```yaml # train w/ shisa-ai/shisa-v1-athenev2-reannotated-filtered base_model: meta-llama/Meta-Llama-3.1-8B-Instruct model_type: LlamaForCausalLM tokenizer_type: AutoTokenizer load_in_8bit: false load_in_4bit: false strict: false # User Liger plugins: - axolotl.integrations.liger.LigerPlugin liger_rope: true liger_rms_norm: true liger_glu_activation: true liger_fused_linear_cross_entropy: true chat_template: llama3 datasets: - path: shisa-ai/shisa-v2-code-math-reasoning-sft-mix type: chat_template field_messages: conversations message_property_mappings: role: role content: content roles: system: - system assistant: - gpt - model - assistant user: - human - user roles_to_train: ["assistant"] dataset_prepared_path: last_run_prepared val_set_size: 0.05 output_dir: ./outputs/ablation-39-cmrmix.masked-shisa-v2-llama-3.1-8b-lr8e6 sequence_len: 8192 sample_packing: true pad_to_sequence_len: true # marginal difference neftune_noise_alpha: 5 use_wandb: true wandb_project: shisa-v2 wandb_entity: augmxnt wandb_name: ablation-39-cmrmix.masked-shisa-v2-llama-3.1-8b-lr8e6 gradient_accumulation_steps: 2 micro_batch_size: 4 num_epochs: 3 optimizer: paged_adamw_8bit lr_scheduler: linear learning_rate: 8e-6 train_on_inputs: false group_by_length: false bf16: auto fp16: tf32: false gradient_checkpointing: true gradient_checkpointing_kwargs: use_reentrant: false early_stopping_patience: resume_from_checkpoint: logging_steps: 1 xformers_attention: flash_attention: true warmup_steps: 100 evals_per_epoch: 2 eval_table_size: saves_per_epoch: 0 save_total_limit: 1 # Only store a single checkpoint debug: deepspeed: zero3_bf16.json weight_decay: 0.00 fsdp: fsdp_config: special_tokens: pad_token: <|end_of_text|> ``` </details><br> # outputs/ablation-39-cmrmix.masked-shisa-v2-llama-3.1-8b-lr8e6 This model is a fine-tuned version of [meta-llama/Meta-Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) on the shisa-ai/shisa-v2-code-math-reasoning-sft-mix dataset. It achieves the following results on the evaluation set: - Loss: 0.4404 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 8e-06 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - distributed_type: multi-GPU - num_devices: 8 - gradient_accumulation_steps: 2 - total_train_batch_size: 64 - total_eval_batch_size: 32 - optimizer: Use OptimizerNames.PAGED_ADAMW_8BIT with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 100 - num_epochs: 3.0 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:------:|:----:|:---------------:| | 0.9144 | 0.0068 | 1 | 0.6696 | | 0.6754 | 0.5 | 74 | 0.4513 | | 0.6119 | 1.0 | 148 | 0.4259 | | 0.5958 | 1.5 | 222 | 0.4243 | | 0.5772 | 2.0 | 296 | 0.4208 | | 0.5074 | 2.5 | 370 | 0.4395 | | 0.4894 | 3.0 | 444 | 0.4404 | ### Framework versions - Transformers 4.48.3 - Pytorch 2.6.0+cu124 - Datasets 3.2.0 - Tokenizers 0.21.0
Xenova/nb-whisper-medium-beta
Xenova
2025-03-07T10:51:59Z
70
0
transformers.js
[ "transformers.js", "onnx", "whisper", "automatic-speech-recognition", "base_model:NbAiLab/nb-whisper-medium-beta", "base_model:quantized:NbAiLab/nb-whisper-medium-beta", "region:us" ]
automatic-speech-recognition
2023-08-29T00:24:06Z
--- base_model: NbAiLab/nb-whisper-medium-beta library_name: transformers.js --- https://huggingface.co/NbAiLab/nb-whisper-medium-beta with ONNX weights to be compatible with Transformers.js. Note: Having a separate repo for ONNX weights is intended to be a temporary solution until WebML gains more traction. If you would like to make your models web-ready, we recommend converting to ONNX using [🤗 Optimum](https://huggingface.co/docs/optimum/index) and structuring your repo like this one (with ONNX weights located in a subfolder named `onnx`).
Xenova/mms-1b-l1107
Xenova
2025-03-07T10:51:43Z
72
0
transformers.js
[ "transformers.js", "onnx", "wav2vec2", "automatic-speech-recognition", "mms", "base_model:facebook/mms-1b-l1107", "base_model:quantized:facebook/mms-1b-l1107", "region:us" ]
automatic-speech-recognition
2023-07-23T17:19:48Z
--- base_model: facebook/mms-1b-l1107 library_name: transformers.js tags: - mms --- https://huggingface.co/facebook/mms-1b-l1107 with ONNX weights to be compatible with Transformers.js. Note: Having a separate repo for ONNX weights is intended to be a temporary solution until WebML gains more traction. If you would like to make your models web-ready, we recommend converting to ONNX using [🤗 Optimum](https://huggingface.co/docs/optimum/index) and structuring your repo like this one (with ONNX weights located in a subfolder named `onnx`).
iFaz/llama32_3B_en_emo_2000_stp
iFaz
2025-03-07T10:51:37Z
0
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "conversational", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "4-bit", "bitsandbytes", "region:us" ]
text-generation
2025-03-07T10:51:11Z
--- base_model: unsloth/llama-3.2-3b-instruct-bnb-4bit tags: - text-generation-inference - transformers - unsloth - llama - trl - sft license: apache-2.0 language: - en --- # Uploaded model - **Developed by:** iFaz - **License:** apache-2.0 - **Finetuned from model :** unsloth/llama-3.2-3b-instruct-bnb-4bit This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
Xenova/beit-large-patch16-512
Xenova
2025-03-07T10:51:32Z
73
0
transformers.js
[ "transformers.js", "onnx", "beit", "image-classification", "base_model:microsoft/beit-large-patch16-512", "base_model:quantized:microsoft/beit-large-patch16-512", "region:us" ]
image-classification
2023-09-05T03:21:22Z
--- base_model: microsoft/beit-large-patch16-512 library_name: transformers.js --- https://huggingface.co/microsoft/beit-large-patch16-512 with ONNX weights to be compatible with Transformers.js. Note: Having a separate repo for ONNX weights is intended to be a temporary solution until WebML gains more traction. If you would like to make your models web-ready, we recommend converting to ONNX using [🤗 Optimum](https://huggingface.co/docs/optimum/index) and structuring your repo like this one (with ONNX weights located in a subfolder named `onnx`).
H1tak3/roberta-lora-phishing_96
H1tak3
2025-03-07T10:50:41Z
0
0
transformers
[ "transformers", "safetensors", "roberta", "text-classification", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-classification
2025-03-07T10:50:21Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
Xenova/w2v-bert-2.0
Xenova
2025-03-07T10:49:43Z
29
1
transformers.js
[ "transformers.js", "onnx", "wav2vec2-bert", "feature-extraction", "base_model:facebook/w2v-bert-2.0", "base_model:quantized:facebook/w2v-bert-2.0", "region:us" ]
feature-extraction
2024-01-26T13:02:08Z
--- base_model: facebook/w2v-bert-2.0 library_name: transformers.js --- https://huggingface.co/facebook/w2v-bert-2.0 with ONNX weights to be compatible with Transformers.js. Note: Having a separate repo for ONNX weights is intended to be a temporary solution until WebML gains more traction. If you would like to make your models web-ready, we recommend converting to ONNX using [🤗 Optimum](https://huggingface.co/docs/optimum/index) and structuring your repo like this one (with ONNX weights located in a subfolder named `onnx`).
cendekiaaa/speecht5_finetuned
cendekiaaa
2025-03-07T10:49:35Z
0
0
transformers
[ "transformers", "tensorboard", "safetensors", "speecht5", "text-to-audio", "generated_from_trainer", "base_model:microsoft/speecht5_tts", "base_model:finetune:microsoft/speecht5_tts", "license:mit", "endpoints_compatible", "region:us" ]
text-to-audio
2025-03-07T10:14:18Z
--- library_name: transformers license: mit base_model: microsoft/speecht5_tts tags: - generated_from_trainer model-index: - name: speecht5_finetuned results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # speecht5_finetuned This model is a fine-tuned version of [microsoft/speecht5_tts](https://huggingface.co/microsoft/speecht5_tts) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.0444 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 4 - eval_batch_size: 2 - seed: 42 - gradient_accumulation_steps: 5 - total_train_batch_size: 20 - optimizer: Use OptimizerNames.ADAMW_TORCH with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 5 - num_epochs: 10 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:------:|:----:|:---------------:| | 2.1694 | 0.5435 | 5 | 1.8182 | | 1.5635 | 1.0 | 10 | 1.4978 | | 1.5795 | 1.5435 | 15 | 1.4283 | | 1.181 | 2.0 | 20 | 1.2522 | | 1.2991 | 2.5435 | 25 | 1.1587 | | 1.1739 | 3.0 | 30 | 1.1261 | | 1.2896 | 3.5435 | 35 | 1.1075 | | 1.0826 | 4.0 | 40 | 1.1085 | | 1.2882 | 4.5435 | 45 | 1.0843 | | 0.9574 | 5.0 | 50 | 1.0844 | | 1.2141 | 5.5435 | 55 | 1.0649 | | 1.1539 | 6.0 | 60 | 1.0672 | | 1.1425 | 6.5435 | 65 | 1.0420 | | 1.0499 | 7.0 | 70 | 1.0359 | | 1.1623 | 7.5435 | 75 | 1.0442 | | 0.9757 | 8.0 | 80 | 1.0372 | | 1.1794 | 8.5435 | 85 | 1.0307 | | 0.9605 | 9.0 | 90 | 1.0444 | ### Framework versions - Transformers 4.48.3 - Pytorch 2.5.1+cu124 - Datasets 3.3.2 - Tokenizers 0.21.0
Xenova/siglip-large-patch16-256
Xenova
2025-03-07T10:49:25Z
24
1
transformers.js
[ "transformers.js", "onnx", "siglip", "zero-shot-image-classification", "base_model:google/siglip-large-patch16-256", "base_model:quantized:google/siglip-large-patch16-256", "region:us" ]
zero-shot-image-classification
2024-01-11T17:58:43Z
--- base_model: google/siglip-large-patch16-256 library_name: transformers.js --- https://huggingface.co/google/siglip-large-patch16-256 with ONNX weights to be compatible with Transformers.js. ## Usage (Transformers.js) If you haven't already, you can install the [Transformers.js](https://huggingface.co/docs/transformers.js) JavaScript library from [NPM](https://www.npmjs.com/package/@huggingface/transformers) using: ```bash npm i @huggingface/transformers ``` **Example:** Zero-shot image classification w/ `Xenova/siglip-large-patch16-256`: ```js import { pipeline } from '@huggingface/transformers'; const classifier = await pipeline('zero-shot-image-classification', 'Xenova/siglip-large-patch16-256'); const url = 'http://images.cocodataset.org/val2017/000000039769.jpg'; const output = await classifier(url, ['2 cats', '2 dogs'], { hypothesis_template: 'a photo of {}', }); console.log(output); // [ // { score: 0.3086719512939453, label: '2 cats' }, // { score: 0.0000623430241830647, label: '2 dogs' } // ] ``` **Example:** Compute text embeddings with `SiglipTextModel`. ```javascript import { AutoTokenizer, SiglipTextModel } from '@huggingface/transformers'; // Load tokenizer and text model const tokenizer = await AutoTokenizer.from_pretrained('Xenova/siglip-large-patch16-256'); const text_model = await SiglipTextModel.from_pretrained('Xenova/siglip-large-patch16-256'); // Run tokenization const texts = ['a photo of 2 cats', 'a photo of 2 dogs']; const text_inputs = tokenizer(texts, { padding: 'max_length', truncation: true }); // Compute embeddings const { pooler_output } = await text_model(text_inputs); // Tensor { // dims: [ 2, 768 ], // type: 'float32', // data: Float32Array(1536) [ ... ], // size: 1536 // } ``` **Example:** Compute vision embeddings with `SiglipVisionModel`. ```javascript import { AutoProcessor, SiglipVisionModel, RawImage } from '@huggingface/transformers'; // Load processor and vision model const processor = await AutoProcessor.from_pretrained('Xenova/siglip-large-patch16-256'); const vision_model = await SiglipVisionModel.from_pretrained('Xenova/siglip-large-patch16-256'); // Read image and run processor const image = await RawImage.read('https://huggingface.co/datasets/Xenova/transformers.js-docs/resolve/main/football-match.jpg'); const image_inputs = await processor(image); // Compute embeddings const { pooler_output } = await vision_model(image_inputs); // Tensor { // dims: [ 1, 768 ], // type: 'float32', // data: Float32Array(768) [ ... ], // size: 768 // } ``` --- Note: Having a separate repo for ONNX weights is intended to be a temporary solution until WebML gains more traction. If you would like to make your models web-ready, we recommend converting to ONNX using [🤗 Optimum](https://huggingface.co/docs/optimum/index) and structuring your repo like this one (with ONNX weights located in a subfolder named `onnx`).
Xenova/siglip-large-patch16-384
Xenova
2025-03-07T10:48:32Z
30
0
transformers.js
[ "transformers.js", "onnx", "siglip", "zero-shot-image-classification", "base_model:google/siglip-large-patch16-384", "base_model:quantized:google/siglip-large-patch16-384", "region:us" ]
zero-shot-image-classification
2024-01-11T18:01:37Z
--- base_model: google/siglip-large-patch16-384 library_name: transformers.js --- https://huggingface.co/google/siglip-large-patch16-384 with ONNX weights to be compatible with Transformers.js. ## Usage (Transformers.js) If you haven't already, you can install the [Transformers.js](https://huggingface.co/docs/transformers.js) JavaScript library from [NPM](https://www.npmjs.com/package/@huggingface/transformers) using: ```bash npm i @huggingface/transformers ``` **Example:** Zero-shot image classification w/ `Xenova/siglip-large-patch16-384`: ```js import { pipeline } from '@huggingface/transformers'; const classifier = await pipeline('zero-shot-image-classification', 'Xenova/siglip-large-patch16-384'); const url = 'http://images.cocodataset.org/val2017/000000039769.jpg'; const output = await classifier(url, ['2 cats', '2 dogs'], { hypothesis_template: 'a photo of {}', }); console.log(output); // [ // { score: 0.4783420264720917, label: '2 cats' }, // { score: 0.00022271279885899276, label: '2 dogs' } // ] ``` **Example:** Compute text embeddings with `SiglipTextModel`. ```javascript import { AutoTokenizer, SiglipTextModel } from '@huggingface/transformers'; // Load tokenizer and text model const tokenizer = await AutoTokenizer.from_pretrained('Xenova/siglip-large-patch16-384'); const text_model = await SiglipTextModel.from_pretrained('Xenova/siglip-large-patch16-384'); // Run tokenization const texts = ['a photo of 2 cats', 'a photo of 2 dogs']; const text_inputs = tokenizer(texts, { padding: 'max_length', truncation: true }); // Compute embeddings const { pooler_output } = await text_model(text_inputs); // Tensor { // dims: [ 2, 768 ], // type: 'float32', // data: Float32Array(1536) [ ... ], // size: 1536 // } ``` **Example:** Compute vision embeddings with `SiglipVisionModel`. ```javascript import { AutoProcessor, SiglipVisionModel, RawImage } from '@huggingface/transformers'; // Load processor and vision model const processor = await AutoProcessor.from_pretrained('Xenova/siglip-large-patch16-384'); const vision_model = await SiglipVisionModel.from_pretrained('Xenova/siglip-large-patch16-384'); // Read image and run processor const image = await RawImage.read('https://huggingface.co/datasets/Xenova/transformers.js-docs/resolve/main/football-match.jpg'); const image_inputs = await processor(image); // Compute embeddings const { pooler_output } = await vision_model(image_inputs); // Tensor { // dims: [ 1, 768 ], // type: 'float32', // data: Float32Array(768) [ ... ], // size: 768 // } ``` --- Note: Having a separate repo for ONNX weights is intended to be a temporary solution until WebML gains more traction. If you would like to make your models web-ready, we recommend converting to ONNX using [🤗 Optimum](https://huggingface.co/docs/optimum/index) and structuring your repo like this one (with ONNX weights located in a subfolder named `onnx`).
devendrah/DeepSeek-R1-Medical-COT
devendrah
2025-03-07T10:48:07Z
0
0
transformers
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "conversational", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
2025-03-07T10:42:54Z
--- base_model: unsloth/deepseek-r1-distill-llama-8b-unsloth-bnb-4bit tags: - text-generation-inference - transformers - unsloth - llama - trl - sft license: apache-2.0 language: - en --- # Uploaded model - **Developed by:** devendrah - **License:** apache-2.0 - **Finetuned from model :** unsloth/deepseek-r1-distill-llama-8b-unsloth-bnb-4bit This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
syntheticbot/ocr-qwen
syntheticbot
2025-03-07T10:46:48Z
44
1
null
[ "safetensors", "qwen2_5_vl", "arxiv:2409.12191", "arxiv:2308.12966", "license:apache-2.0", "region:us" ]
null
2025-02-17T07:08:28Z
--- license: apache-2.0 --- # syntheticbot/ocr-qwen ## Introduction syntheticbot/ocr-qwen is a fine-tuned model for Optical Character Recognition (OCR) tasks, derived from the base model [Qwen/Qwen2.5-VL-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-VL-7B-Instruct). This model is engineered for high accuracy in extracting text from images, including documents and scenes containing text. #### Key Enhancements for OCR: * **Enhanced Text Recognition Accuracy**: Superior accuracy across diverse text fonts, styles, sizes, and orientations. * **Robustness to Document Variations**: Specifically trained to manage document complexities like varied layouts, noise, and distortions. * **Structured Output Generation**: Enables structured output formats (JSON, CSV) for recognized text and layout in document images such as invoices and tables. * **Text Localization**: Provides accurate localization of text regions and bounding boxes for text elements within images. * **Improved Handling of Text in Visuals**: Maintains proficiency in analyzing charts and graphics, with enhanced recognition of embedded text. #### Model Architecture Updates: * **Dynamic Resolution and Frame Rate Training for Video Understanding**: <p align="center"> <img src="https://qianwen-res.oss-cn-beijing.aliyuncs.com/Qwen2.5-VL/qwen2.5vl_arc.jpeg" width="80%"/> <p> * **Streamlined and Efficient Vision Encoder** This repository provides the instruction-tuned and OCR-optimized 7B Qwen-VL-7B-ocr model. For comprehensive details about the foundational model architecture, please refer to the [Qwen/Qwen2.5-VL-7B-Instruct](https://huggingface.co/Qwen/Qwen2.5-VL-7B-Instruct) repository, as well as the [Blog](https://qwenlm.github.io/blog/qwen2.5-vl/) and [GitHub](https://github.com/QwenLM/Qwen2.5-VL) pages for Qwen2.5-VL. ## Requirements For optimal performance and access to OCR-specific features, it is recommended to build from source: ``` pip install git+https://github.com/huggingface/transformers accelerate ``` ## Quickstart The following examples illustrate the use of syntheticbot/ocr-qwen with 🤗 Transformers and `qwen_vl_utils` for OCR applications. ``` pip install git+https://github.com/huggingface/transformers accelerate ``` Install the toolkit for streamlined visual input processing: ```bash pip install qwen-vl-utils[decord]==0.0.8 ``` ### Using 🤗 Transformers for OCR ```python from transformers import Qwen2_5_VLForConditionalGeneration, AutoProcessor from qwen_vl_utils import process_vision_info import torch model = Qwen2_5_VLForConditionalGeneration.from_pretrained( "syntheticbot/ocr-qwen", torch_dtype="auto", device_map="auto" ) processor = AutoProcessor.from_pretrained("syntheticbot/ocr-qwen") messages = [ { "role": "user", "content": [ { "type": "image", "image": "path/to/your/document_image.jpg", }, {"type": "text", "text": "Extract the text from this image."}, ], } ] text = processor.apply_chat_template( messages, tokenize=False, add_generation_prompt=True ) image_inputs, video_inputs = process_vision_info(messages) inputs = processor( text=[text], images=image_inputs, videos=video_inputs, padding=True, return_tensors="pt", ) inputs = inputs.to("cuda" if torch.cuda.is_available() else "cpu") generated_ids = model.generate(**inputs, max_new_tokens=512) generated_ids_trimmed = [ out_ids[len(in_ids) :] for in_ids, out_ids in zip(inputs.input_ids, generated_ids) ] output_text = processor.batch_decode( generated_ids_trimmed, skip_special_tokens=True, clean_up_tokenization_spaces=False ) print("Extracted Text:", output_text[0]) ``` <details> <summary>Example for Structured Output (JSON for Table Extraction)</summary> ```python from transformers import Qwen2_5_VLForConditionalGeneration, AutoProcessor from qwen_vl_utils import process_vision_info import torch import json model = Qwen2_5_VLForConditionalGeneration.from_pretrained( "syntheticbot/ocr-qwen", torch_dtype="auto", device_map="auto" ) processor = AutoProcessor.from_pretrained("syntheticbot/ocr-qwen") messages = [ { "role": "user", "content": [ { "type": "image", "image": "path/to/your/table_image.jpg", }, {"type": "text", "text": "Extract the table from this image and output it as JSON."}, ], } ] text = processor.apply_chat_template( messages, tokenize=False, add_generation_prompt=True ) image_inputs, video_inputs = process_vision_info(messages) inputs = processor( text=[text], images=image_inputs, videos=video_inputs, padding=True, return_tensors="pt", ) inputs = inputs.to("cuda" if torch.cuda.is_available() else "cpu") generated_ids = model.generate(**inputs, max_new_tokens=1024) generated_ids_trimmed = [ out_ids[len(in_ids) :] for in_ids, out_ids in zip(inputs.input_ids, generated_ids) ] output_text = processor.batch_decode( generated_ids_trimmed, skip_special_tokens=True, clean_up_tokenization_spaces=False ) print("Extracted Table (JSON):\n", output_text[0]) try: json_output = json.loads(output_text[0]) print("\nParsed JSON Output:\n", json.dumps(json_output, indent=2)) except json.JSONDecodeError: print("\nCould not parse output as JSON. Output is plain text.") ``` </details> <details> <summary>Batch inference for OCR</summary> ```python messages1 = [ { "role": "user", "content": [ {"type": "image", "image": "path/to/image1.jpg"}, {"type": "text", "text": "Extract text from this image."}, ], } ] messages2 = [ { "role": "user", "content": [ {"type": "image", "image": "path/to/image2.jpg"}, {"type": "text", "text": "Read the text in this document."}, ], } ] messages = [messages1, messages2] texts = [ processor.apply_chat_template(msg, tokenize=False, add_generation_prompt=True) for msg in messages ] image_inputs, video_inputs = process_vision_info(messages) inputs = processor( text=texts, images=image_inputs, videos=video_inputs, padding=True, return_tensors="pt", ) inputs = inputs.to("cuda" if torch.cuda.is_available() else "cpu") generated_ids = model.generate(**inputs, max_new_tokens=512) generated_ids_trimmed = [ out_ids[len(in_ids) :] for in_ids, out_ids in zip(inputs.input_ids, generated_ids) ] output_texts = processor.batch_decode( generated_ids_trimmed, skip_special_tokens=True, clean_up_tokenization_spaces=False ) print("Extracted Texts (Batch):\n", output_texts) ``` </details> ### 🤖 ModelScope For users in mainland China, ModelScope is recommended. Use `snapshot_download` for checkpoint management. Adapt model names to `syntheticbot/ocr-qwen` in ModelScope implementations. ### More Usage Tips for OCR Input images support local files, URLs, and base64 encoding. ```python messages = [ { "role": "user", "content": [ { "type": "image", "image": "http://path/to/your/document_image.jpg" }, { "type": "text", "text": "Extract the text from this image URL." }, ], } ] ``` #### Image Resolution for OCR Accuracy Higher resolution images typically improve OCR accuracy, especially for small text. Adjust resolution using `min_pixels`, `max_pixels`, `resized_height`, and `resized_width` parameters with the processor. ```python min_pixels = 512 * 28 * 28 max_pixels = 2048 * 28 * 28 processor = AutoProcessor.from_pretrained( "syntheticbot/ocr-qwen", min_pixels=min_pixels, max_pixels=max_pixels ) ``` Control resizing dimensions directly: ```python messages = [ { "role": "user", "content": [ { "type": "image", "image": "file:///path/to/your/document_image.jpg", "resized_height": 600, "resized_width": 800, }, {"type": "text", "text": "Extract the text."}, ], } ] ``` ## Citation from base model ``` @misc{qwen2.5-VL, title = {Qwen2.5-VL}, url = {https://qwenlm.github.io/blog/qwen2.5-vl/}, author = {Qwen Team}, month = {January}, year = {2025} } @article{Qwen2VL, title={Qwen2-VL: Enhancing Vision-Language Model's Perception of the World at Any Resolution}, author={Wang, Peng and Bai, Shuai and Tan, Sinan and Wang, Shijie and Fan, Zhihao and Bai, Jinze and Chen, Keqin and Liu, Xuejing and Wang, Jialin and Ge, Wenbin and Fan, Yang and Dang, Kai and Du, Mengfei and Ren, Xuancheng and Men, Rui and Liu, Dayiheng and Zhou, Chang and Zhou, Jingren and Lin, Junyang}, journal={arXiv preprint arXiv:2409.12191}, year={2024} } @article{Qwen-VL, title={Qwen-VL: A Versatile Vision-Language Model for Understanding, Localization, Text Reading, and Beyond}, author={Bai, Jinze and Bai, Shuai and Yang, Shusheng and Wang, Shijie and Tan, Sinan and Wang, Peng and Lin, Junyang and Zhou, Chang and Zhou, Jingren}, journal={arXiv preprint arXiv:2308.12966}, year={2023} } ```
baby-dev/a7ad267e-2b01-4c44-98ed-af0dbcec558b
baby-dev
2025-03-07T10:44:44Z
0
0
peft
[ "peft", "generated_from_trainer", "base_model:fxmarty/really-tiny-falcon-testing", "base_model:adapter:fxmarty/really-tiny-falcon-testing", "region:us" ]
null
2025-03-07T10:44:38Z
--- library_name: peft tags: - generated_from_trainer base_model: fxmarty/really-tiny-falcon-testing model-index: - name: baby-dev/a7ad267e-2b01-4c44-98ed-af0dbcec558b results: [] --- <!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # baby-dev/a7ad267e-2b01-4c44-98ed-af0dbcec558b This model was trained from scratch on the None dataset. It achieves the following results on the evaluation set: - Loss: 10.8551 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ### Framework versions - PEFT 0.13.2 - Transformers 4.46.0 - Pytorch 2.5.0+cu124 - Datasets 3.0.1 - Tokenizers 0.20.1
MeiKing111/SN09_COM4_68
MeiKing111
2025-03-07T10:44:14Z
0
0
transformers
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
2025-03-07T09:30:26Z
--- library_name: transformers tags: [] --- # Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
Xenova/sam-vit-huge
Xenova
2025-03-07T10:43:54Z
42
1
transformers.js
[ "transformers.js", "onnx", "sam", "mask-generation", "base_model:facebook/sam-vit-huge", "base_model:quantized:facebook/sam-vit-huge", "region:us" ]
mask-generation
2023-05-31T11:54:25Z
--- base_model: facebook/sam-vit-huge library_name: transformers.js --- https://huggingface.co/facebook/sam-vit-huge with ONNX weights to be compatible with Transformers.js. ## Usage (Transformers.js) If you haven't already, you can install the [Transformers.js](https://huggingface.co/docs/transformers.js) JavaScript library from [NPM](https://www.npmjs.com/package/@huggingface/transformers) using: ```bash npm i @huggingface/transformers ``` **Example:** Perform mask generation with `Xenova/sam-vit-huge`. ```js import { SamModel, AutoProcessor, RawImage } from "@huggingface/transformers"; // Load model and processor const model = await SamModel.from_pretrained("Xenova/sam-vit-huge"); const processor = await AutoProcessor.from_pretrained("Xenova/sam-vit-huge"); // Prepare image and input points const img_url = "https://huggingface.co/datasets/Xenova/transformers.js-docs/resolve/main/corgi.jpg"; const raw_image = await RawImage.read(img_url); const input_points = [[[340, 250]]]; // Process inputs and perform mask generation const inputs = await processor(raw_image, { input_points }); const outputs = await model(inputs); // Post-process masks const masks = await processor.post_process_masks(outputs.pred_masks, inputs.original_sizes, inputs.reshaped_input_sizes); console.log(masks); // [ // Tensor { // dims: [ 1, 3, 410, 614 ], // type: 'bool', // data: Uint8Array(755220) [ ... ], // size: 755220 // } // ] const scores = outputs.iou_scores; console.log(scores); // Tensor { // dims: [ 1, 1, 3 ], // type: 'float32', // data: Float32Array(3) [ // 0.9742214679718018, // 1.002995491027832, // 0.9613651037216187 // ], // size: 3 // } ``` You can then visualize the generated mask with: ```js const image = RawImage.fromTensor(masks[0][0].mul(255)); image.save('mask.png'); ``` ![image/png](https://cdn-uploads.huggingface.co/production/uploads/61b253b7ac5ecaae3d1efe0c/naMfUnwlCZxPkpbe7nvzQ.png) Next, select the channel with the highest IoU score, which in this case is the second (green) channel. Intersecting this with the original image gives us an isolated version of the subject: ![image/gif](https://cdn-uploads.huggingface.co/production/uploads/61b253b7ac5ecaae3d1efe0c/uuNUEp7K_GaiYWMbk_x29.gif) ## Demo We've also got an online demo, which you can try out [here](https://huggingface.co/spaces/Xenova/segment-anything-web). <video controls autoplay src="https://cdn-uploads.huggingface.co/production/uploads/61b253b7ac5ecaae3d1efe0c/Y0wAOw6hz9rWpwiuMoz2A.mp4"></video> --- Note: Having a separate repo for ONNX weights is intended to be a temporary solution until WebML gains more traction. If you would like to make your models web-ready, we recommend converting to ONNX using [🤗 Optimum](https://huggingface.co/docs/optimum/index) and structuring your repo like this one (with ONNX weights located in a subfolder named `onnx`).
seongil-dn/bge-m3-831
seongil-dn
2025-03-07T10:43:50Z
0
0
sentence-transformers
[ "sentence-transformers", "safetensors", "xlm-roberta", "sentence-similarity", "feature-extraction", "generated_from_trainer", "dataset_size:1138596", "loss:CachedGISTEmbedLoss", "arxiv:1908.10084", "base_model:seongil-dn/unsupervised_20m_3800", "base_model:finetune:seongil-dn/unsupervised_20m_3800", "autotrain_compatible", "text-embeddings-inference", "endpoints_compatible", "region:us" ]
sentence-similarity
2025-03-07T10:39:35Z
--- tags: - sentence-transformers - sentence-similarity - feature-extraction - generated_from_trainer - dataset_size:1138596 - loss:CachedGISTEmbedLoss base_model: seongil-dn/unsupervised_20m_3800 widget: - source_sentence: How many people were reported to have died in the Great Fire of London in 1666? sentences: - City of London 1666. Both of these fires were referred to as "the" Great Fire. After the fire of 1666, a number of plans were drawn up to remodel the City and its street pattern into a renaissance-style city with planned urban blocks, squares and boulevards. These plans were almost entirely not taken up, and the medieval street pattern re-emerged almost intact. By the late 16th century, London increasingly became a major centre for banking, international trade and commerce. The Royal Exchange was founded in 1565 by Sir Thomas Gresham as a centre of commerce for London's merchants, and gained Royal patronage in - Great Atlanta fire of 1917 Great Atlanta fire of 1917 The Great Atlanta Fire of 1917 began just after noon on 21 May 1917 in the Old Fourth Ward of Atlanta, Georgia. It is unclear just how the fire started, but it was fueled by hot temperatures and strong winds which propelled the fire. The fire, which burned for nearly 10 hours, destroyed and 1,900 structures displacing over 10,000 people. Damages were estimated at $5 million, ($ million when adjusted for inflation). It was a clear, warm and sunny day with a brisk breeze from the south. This was not the only fire of the - Great Plague of London they had ever been seen ...". Plague cases continued to occur sporadically at a modest rate until the summer of 1666. On the second and third of September that year, the Great Fire of London destroyed much of the City of London, and some people believed that the fire put an end to the epidemic. However, it is now thought that the plague had largely subsided before the fire took place. In fact, most of the later cases of plague were found in the suburbs, and it was the City of London itself that was destroyed by the Fire. According - Monument to the Great Fire of London Monument to the Great Fire of London The Monument to the Great Fire of London, more commonly known simply as the Monument, is a Doric column in London, United Kingdom, situated near the northern end of London Bridge. Commemorating the Great Fire of London, it stands at the junction of Monument Street and Fish Street Hill, in height and 202 feet west of the spot in Pudding Lane where the Great Fire started on 2 September 1666. Constructed between 1671 and 1677, it was built on the site of St. Margaret's, Fish Street, the first church to be destroyed by - 'How to Have Sex in an Epidemic New York City government and organizations within the LGBT community. The Gay Men''s Health Crisis offered to buy all 5,000 pamphlets and promote them, with the condition that any mentions of the multifactorial model be removed from the writing. The authors refused. Berkowitz recounts in an interview it being "infuriating" that in 1985, the city still hadn''t adopted any standard safe sex education. The advent of safe sex in urban gay male populations came too late for many people: by 1983, more than 1,476 people had died from AIDS and David France estimated that as much as half of all' - 'Monument to the Great Fire of London six years to complete the 202 ft column. It was two more years before the inscription (which had been left to Wren — or to Wren''s choice — to decide upon) was set in place. "Commemorating — with a brazen disregard for the truth — the fact that ''London rises again...three short years complete that which was considered the work of ages.''" Hooke''s surviving drawings show that several versions of the monument were submitted for consideration: a plain obelisk, a column garnished with tongues of fire, and the fluted Doric column that was eventually chosen. The real contention came with' - source_sentence: '"The Claude Francois song ""Comme d''habitude"" (translation ""as usual"") was a hit in English for Frank Sinatra under what title?"' sentences: - Young at Heart (Frank Sinatra song) young, Dick Van Dyke recorded a duet with his wife, Arlene, at Capital Records Studio in Los Angeles, filmed for the HBO Special on aging "If I'm not in the Obituary, I'll have Breakfast" starring Carl Reiner, and featuring other young at heart +90 treasures, Mel Brooks, Norman Lear, Stan Lee & Betty White among others. Van Dyke was recorded using Frank Sinatra's microphone. Young at Heart (Frank Sinatra song) "Young at Heart" is a pop standard, a ballad with music by Johnny Richards and lyrics by Carolyn Leigh. The song was written and published in 1953, with Leigh contributing - 'Comme d''habitude a relationship that is falling out of love, while the English language version is set at the end of a lifetime, approaching death, and looking back without regret – expressing feelings that are more related to Piaf''s song "Non, je ne regrette rien". Many artists sang "Comme d''Habitude" in French after Claude François''s success (and international success through ''"My Way"), notably: David Bowie has said that in 1968 – the year before Paul Anka acquired the French song – his manager, Kenneth Pitt, asked him to write English lyrics for "Comme d''habitude" but that his version, titled "Even a Fool' - Frank Sinatra Me" with Billy May, designed as a musical world tour. It reached the top spot on the Billboard album chart in its second week, remaining at the top for five weeks, and was nominated for the Grammy Award for Album of the Year at the inaugural Grammy Awards. The title song, "Come Fly With Me", written especially for him, would become one of his best known standards. On May 29 he recorded seven songs in a single session, more than double the usual yield of a recording session, and an eighth was planned, "Lush Life", but Sinatra found it too - Frank Sinatra Original Song. Sinatra released "Softly, as I Leave You", and collaborated with Bing Crosby and Fred Waring on "America, I Hear You Singing", a collection of patriotic songs recorded as a tribute to the assassinated President John F. Kennedy. Sinatra increasingly became involved in charitable pursuits in this period. In 1961 and 1962 he went to Mexico, with the sole purpose of putting on performances for Mexican charities, and in July 1964 he was present for the dedication of the Frank Sinatra International Youth Center for Arab and Jewish children in Nazareth. Sinatra's phenomenal success in 1965, coinciding with his - Comme ci comme ça (Basim song) to the charm of it all. Working both Danish and Moroccan Arabic, Basim sings about a girl he is ready to commit to. It doesn’t mater what she wants to do — it’s comme ci comme ça — and he just wants her." An official music video to accompany the release of "Comme ci comme ça" was first released onto YouTube on 20 September 2017 at a total length of three minutes and twelve seconds. Comme ci comme ça (Basim song) "Comme ci comme ça" is a song performed by Danish pop singer and songwriter Basim, featuring vocals from Gilli. - Personal life of Frank Sinatra A third child, Christina Sinatra, known as "Tina", was born on June 20, 1948. Nancy Barbato Sinatra and Frank Sinatra announced their separation on Valentine's Day, February 14, 1950, with Frank's additional extra-marital affair with Ava Gardner compounding his transgressions and becoming public knowledge once again. After originally just seeking a legal separation, Frank and Nancy Sinatra decided some months later to file for divorce, and this divorce became legally final on October 29, 1951. Frank Sinatra's affair and relationship with Gardner had become more and more serious, and she later became his second wife. What was perhaps less widely - source_sentence: What was the name of the first Indiana Jones movie? sentences: - Indiana Jones and the Temple of Doom point. Old-time, 15-part movie serials didn't have shape. They just went on and on and on, which is what "Temple of Doom" does with humor and technical invention." Neal Gabler commented that "I think in some ways, "Indiana Jones and the Temple of Doom" was better than "Raiders of the Lost Ark". In some ways it was less. In sum total, I'd have to say I enjoyed it more. That doesn't mean it's better necessarily, but I got more enjoyment out of it." Colin Covert of the "Star Tribune" called the film "sillier, darkly violent and a bit dumbed down, - Indiana Jones and the Temple of Doom (1985 video game) Theme music plays in the background which is the best part of the game. Most of the sound effects are not sharp and not enough of them exist. "Indiana Jones and the Temple of Doom" is a bad game all the way around. It looks bad, has bad controls, and is way too short." Indiana Jones and the Temple of Doom (1985 video game) Indiana Jones and The Temple of Doom is a 1985 action arcade game developed and published by Atari Games, based on the 1984 film of the same name, the second film in the "Indiana Jones" franchise. - Indiana Jones and the Spear of Destiny Indiana Jones and the Spear of Destiny Indiana Jones and The Spear of Destiny is a four-issue comic book mini-series published by Dark Horse Comics from April to July 1995. It was their seventh series about the adult Indiana Jones. Indiana Jones reached for the Holy Grail, perched in a crack in the Temple of the Sun. Hanging onto him, his father, Professor Henry Jones urged him to let it go, and Indy turned back and let his father help him up. As the Joneses ride out into the Canyon of the Crescent Moon with Marcus Brody and Sallah, they - Lego Indiana Jones sets" The line was discontinued in 2010, but since Lucas plans to make a fifth installment to the franchise, the sets may be re-released along with new sets of the possible fifth Indiana Jones film. Due to the fact Disney bought Lucasfilm and will be making a new Indiana Jones movie, chances of new sets are high. The Indiana Jones sets proved to be one of the most popular Lego themes, and by the end of 2008 were credited, along with Lego Star Wars, of boosting the Lego Group's profits within a stagnant toy market. The product line was said - Indiana Jones and the Staff of Kings point-and-click adventure "Indiana Jones and the Fate of Atlantis". GameSpot criticized its "terribly laid-out checkpoints", "out-of-date" visuals, and "atrocious, annoying motion controls". Indiana Jones and the Staff of Kings The game was initially developed for the higher-end PlayStation 3 and Xbox 360 systems, before switching to the aforementioned lower-end platforms. As a result, both systems never saw a proper "Indiana Jones" video game being released besides the "" duology. The plot centers around Indy's search for the Staff of Moses. The Wii version of the game includes an exclusive co-op story mode (with Indy and Henry Jones Sr.) and unlockable - 'Indiana Jones and the Last Crusade: The Graphic Adventure Indiana Jones and the Last Crusade: The Graphic Adventure Indiana Jones and the Last Crusade: The Graphic Adventure is a graphic adventure game, released in 1989 (to coincide with the release of the film of the same name), published by Lucasfilm Games (now LucasArts). It was the third game to use the SCUMM engine. "Last Crusade" was one of the most innovative of the LucasArts adventures. It expanded on LucasArts'' traditional adventure game structure by including a flexible point system—the IQ score, or "Indy Quotient"—and by allowing the game to be completed in several different ways. The point system was' - source_sentence: '"Who was the Anglo-Irish scientist who, in the 17th century, discovered that ""the volume of a given mass of gas at a given temperature is inversely proportional to its pressure""?"' sentences: - 'Gay-Lussac''s law Gay-Lussac''s law Gay-Lussac''s law can refer to several discoveries made by French chemist Joseph Louis Gay-Lussac (1778–1850) and other scientists in the late 18th and early 19th centuries pertaining to thermal expansion of gases and the relationship between temperature, volume, and pressure. It states that the pressure of a given mass of gas varies directly with the absolute temperature of the gas, when the volume is kept constant. Mathematically, it can be written as: P/T=constant, Gay-Lussac is most often recognized for the Pressure Law which established that the pressure of an enclosed gas is directly proportional to its temperature and' - 'Gas constant "V" is the volume of gas (SI unit cubic metres), "n" is the amount of gas (SI unit moles), "m" is the mass (SI unit kilograms) contained in "V", and "T" is the thermodynamic temperature (SI unit kelvins). "R" is the molar-weight-specific gas constant, discussed below. The gas constant is expressed in the same physical units as molar entropy and molar heat capacity. From the general equation "PV" = "nRT" we get: where "P" is pressure, "V" is volume, "n" is number of moles of a given substance, and "T" is temperature. As pressure is defined as force per unit' - The Boy Who Was a King term. The film presents not only the life of the former Tsar, but also intertwines within the story vignettes of various Bulgarians, who were supporting him, sending him gifts, or merely tattooing his face on their body. The story is told through personal footage and vast amounts of archive material. The film received praise for its editing and use of archives with Variety's Robert Koehler writing that "Pic’s terrific use of archival footage includes an exiled Simeon interviewed in the early ’60s, disputing his playboy rep." and "Editing is aces." The Boy Who Was a King The Boy Who Was - Francis Hauksbee In 1708, Hauksbee independently discovered Charles's law of gases, which states that, for a given mass of gas at a constant pressure, the volume of the gas is proportional to its temperature. Hauksbee published accounts of his experiments in the Royal Society's journal "Philosophical Transactions". In 1709 he self-published "Physico-Mechanical Experiments on Various Subjects" which collected together many of these experiments along with discussion that summarized much of his scientific work. An Italian translation was published in 1716. A second edition was published posthumously in 1719. There were also translations to Dutch (1735) and French (1754). The Royal Society Hauksbee - 'Boyle''s law air moves from high to low pressure. Related phenomena: Other gas laws: Boyle''s law Boyle''s law, sometimes referred to as the Boyle–Mariotte law, or Mariotte''s law (especially in France), is an experimental gas law that describes how the pressure of a gas tends to increase as the volume of the container decreases. A modern statement of Boyle''s law is The absolute pressure exerted by a given mass of an ideal gas is inversely proportional to the volume it occupies if the temperature and amount of gas remain unchanged within a closed system. Mathematically, Boyle''s law can be stated as or' - Boyle's law of the gas, and "k" is a constant. The equation states that the product of pressure and volume is a constant for a given mass of confined gas and this holds as long as the temperature is constant. For comparing the same substance under two different sets of conditions, the law can be usefully expressed as The equation shows that, as volume increases, the pressure of the gas decreases in proportion. Similarly, as volume decreases, the pressure of the gas increases. The law was named after chemist and physicist Robert Boyle, who published the original law in 1662. This relationship - source_sentence: Peter Stuyvesant, born in Holland, became Governor of which American city in 1647? sentences: - Peter Stuyvesant at the corner of Thirteenth Street and Third Avenue until 1867 when it was destroyed by a storm, bearing fruit almost to the last. The house was destroyed by fire in 1777. He also built an executive mansion of stone called Whitehall. In 1645, Stuyvesant married Judith Bayard (–1687) of the Bayard family. Her brother, Samuel Bayard, was the husband of Stuyvesant's sister, Anna Stuyvesant. Petrus and Judith had two sons together. He died in August 1672 and his body was entombed in the east wall of St. Mark's Church in-the-Bowery, which sits on the site of Stuyvesant’s family chapel. - 'Peter Stuyvesant (cigarette) can amount to millions of dollars and finally criminal prosecution - if companies wilfully break the laws. However last year, when questioned on why no such action was being pursued against Imperial Tobacco a spokeswoman for Federal Health said: ""No instances of non-compliance with the Act have been identified by the Department that warrant the initiation of Court proceedings in the first instance, and without attempting alternative dispute resolution to achieve compliance"". Peter Stuyvesant is or was sold in the following countries: Canada, United States, United Kingdom, Luxembourg, Belgium, The Netherlands, Germany, France, Austria, Switzerland, Spain, Italy, Czech Republic, Greece,' - Jochem Pietersen Kuyter September 25, 1647, until the city was incorporated, in 1653, when he was made schout (sheriff). Kuyter twice came in conflict with the Director of New Netherland. Kuyter was a man of good education, what is evident by his dealings with Willem Kieft., who he believed damaged the colony with his policies and the start of Kieft's War in 1643. In 1647, when Peter Stuyvesant arrived in New Amsterdam to replace Kieft, Kuyter and Cornelis Melyn acting in name of the citizens of New Amsterdam, brought charges against the outgoing governor, demanding an investigation of his conduct while in office. - Peter Stuyvesant (cigarette) half of its regular users"" and called the packaging changes ""the ultimate sick joke from big tobacco"". In 2013, it was reported that Imperial Tobacco Australia had sent marketing material to WA tobacco retailers which promotes limited edition packs of "Peter Stuyvesant + Loosie", which came with 26 cigarettes. The material included images of a young woman with pink hair putting on lipstick and men on the streets of New York and also included a calendar and small poster that were clearly intended to glamorise smoking. Anti-smoking campaigner Mike Daube said although the material did not break the law because - 'Peter Stuyvesant but the order was soon revoked under pressure from the States of Holland and the city of Amsterdam. Stuyvesant prepared against an attack by ordering the citizens to dig a ditch from the North River to the East River and to erect a fortification. In 1653, a convention of two deputies from each village in New Netherland demanded reforms, and Stuyvesant commanded that assembly to disperse, saying: "We derive our authority from God and the company, not from a few ignorant subjects." In the summer of 1655, he sailed down the Delaware River with a fleet of seven vessels and' - Peter Stuyvesant Dutch Reformed church, a Calvinist denomination, holding to the Three Forms of Unity (Belgic Confession, Heidelberg Catechism, Canons of Dordt). The English were Anglicans, holding to the 39 Articles, a Protestant confession, with bishops. In 1665, Stuyvesant went to the Netherlands to report on his term as governor. On his return to the colony, he spent the remainder of his life on his farm of sixty-two acres outside the city, called the Great Bouwerie, beyond which stretched the woods and swamps of the village of Nieuw Haarlem. A pear tree that he reputedly brought from the Netherlands in 1647 remained pipeline_tag: sentence-similarity library_name: sentence-transformers --- # SentenceTransformer based on seongil-dn/unsupervised_20m_3800 This is a [sentence-transformers](https://www.SBERT.net) model finetuned from [seongil-dn/unsupervised_20m_3800](https://huggingface.co/seongil-dn/unsupervised_20m_3800). It maps sentences & paragraphs to a 1024-dimensional dense vector space and can be used for semantic textual similarity, semantic search, paraphrase mining, text classification, clustering, and more. ## Model Details ### Model Description - **Model Type:** Sentence Transformer - **Base model:** [seongil-dn/unsupervised_20m_3800](https://huggingface.co/seongil-dn/unsupervised_20m_3800) <!-- at revision 1cda749f242e2b5c9e4f3c1122a61e76fec1fee5 --> - **Maximum Sequence Length:** 1024 tokens - **Output Dimensionality:** 1024 dimensions - **Similarity Function:** Cosine Similarity <!-- - **Training Dataset:** Unknown --> <!-- - **Language:** Unknown --> <!-- - **License:** Unknown --> ### Model Sources - **Documentation:** [Sentence Transformers Documentation](https://sbert.net) - **Repository:** [Sentence Transformers on GitHub](https://github.com/UKPLab/sentence-transformers) - **Hugging Face:** [Sentence Transformers on Hugging Face](https://huggingface.co/models?library=sentence-transformers) ### Full Model Architecture ``` SentenceTransformer( (0): Transformer({'max_seq_length': 1024, 'do_lower_case': False}) with Transformer model: XLMRobertaModel (1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True}) (2): Normalize() ) ``` ## Usage ### Direct Usage (Sentence Transformers) First install the Sentence Transformers library: ```bash pip install -U sentence-transformers ``` Then you can load this model and run inference. ```python from sentence_transformers import SentenceTransformer # Download from the 🤗 Hub model = SentenceTransformer("seongil-dn/bge-m3-831") # Run inference sentences = [ 'Peter Stuyvesant, born in Holland, became Governor of which American city in 1647?', 'Peter Stuyvesant (cigarette) half of its regular users"" and called the packaging changes ""the ultimate sick joke from big tobacco"". In 2013, it was reported that Imperial Tobacco Australia had sent marketing material to WA tobacco retailers which promotes limited edition packs of "Peter Stuyvesant + Loosie", which came with 26 cigarettes. The material included images of a young woman with pink hair putting on lipstick and men on the streets of New York and also included a calendar and small poster that were clearly intended to glamorise smoking. Anti-smoking campaigner Mike Daube said although the material did not break the law because', 'Peter Stuyvesant (cigarette) can amount to millions of dollars and finally criminal prosecution - if companies wilfully break the laws. However last year, when questioned on why no such action was being pursued against Imperial Tobacco a spokeswoman for Federal Health said: ""No instances of non-compliance with the Act have been identified by the Department that warrant the initiation of Court proceedings in the first instance, and without attempting alternative dispute resolution to achieve compliance"". Peter Stuyvesant is or was sold in the following countries: Canada, United States, United Kingdom, Luxembourg, Belgium, The Netherlands, Germany, France, Austria, Switzerland, Spain, Italy, Czech Republic, Greece,', ] embeddings = model.encode(sentences) print(embeddings.shape) # [3, 1024] # Get the similarity scores for the embeddings similarities = model.similarity(embeddings, embeddings) print(similarities.shape) # [3, 3] ``` <!-- ### Direct Usage (Transformers) <details><summary>Click to see the direct usage in Transformers</summary> </details> --> <!-- ### Downstream Usage (Sentence Transformers) You can finetune this model on your own dataset. <details><summary>Click to expand</summary> </details> --> <!-- ### Out-of-Scope Use *List how the model may foreseeably be misused and address what users ought not to do with the model.* --> <!-- ## Bias, Risks and Limitations *What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.* --> <!-- ### Recommendations *What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.* --> ## Training Details ### Training Dataset #### Unnamed Dataset * Size: 1,138,596 training samples * Columns: <code>anchor</code>, <code>positive</code>, <code>negative</code>, <code>negative_2</code>, <code>negative_3</code>, <code>negative_4</code>, and <code>negative_5</code> * Approximate statistics based on the first 1000 samples: | | anchor | positive | negative | negative_2 | negative_3 | negative_4 | negative_5 | |:--------|:-----------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------|:--------------------------------------------------------------------------------------| | type | string | string | string | string | string | string | string | | details | <ul><li>min: 9 tokens</li><li>mean: 22.32 tokens</li><li>max: 119 tokens</li></ul> | <ul><li>min: 127 tokens</li><li>mean: 157.45 tokens</li><li>max: 420 tokens</li></ul> | <ul><li>min: 122 tokens</li><li>mean: 154.65 tokens</li><li>max: 212 tokens</li></ul> | <ul><li>min: 122 tokens</li><li>mean: 155.52 tokens</li><li>max: 218 tokens</li></ul> | <ul><li>min: 122 tokens</li><li>mean: 156.04 tokens</li><li>max: 284 tokens</li></ul> | <ul><li>min: 124 tokens</li><li>mean: 156.3 tokens</li><li>max: 268 tokens</li></ul> | <ul><li>min: 121 tokens</li><li>mean: 156.15 tokens</li><li>max: 249 tokens</li></ul> | * Samples: | anchor | positive | negative | negative_2 | negative_3 | negative_4 | negative_5 | |:---------------------------------------------------------------------------------------------------------|:-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:-------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------| | <code>What African country is projected to pass the United States in population by the year 2055?</code> | <code>African immigration to the United States officially 40,000 African immigrants, although it has been estimated that the population is actually four times this number when considering undocumented immigrants. The majority of these immigrants were born in Ethiopia, Egypt, Nigeria, and South Africa. African immigrants like many other immigrant groups are likely to establish and find success in small businesses. Many Africans that have seen the social and economic stability that comes from ethnic enclaves such as Chinatowns have recently been establishing ethnic enclaves of their own at much higher rates to reap the benefits of such communities. Such examples include Little Ethiopia in Los Angeles and</code> | <code>What Will Happen to the Gang Next Year? watching television at the time of the broadcast. This made it the lowest-rated episode in "30 Rock"<nowiki>'</nowiki>s history. and a decrease from the previous episode "The Return of Avery Jessup" (2.92 million) What Will Happen to the Gang Next Year? "What Will Happen to the Gang Next Year?" is the twenty-second and final episode of the sixth season of the American television comedy series "30 Rock", and the 125th overall episode of the series. It was directed by Michael Engler, and written by Matt Hubbard. The episode originally aired on the National Broadcasting Company (NBC) network in the United States</code> | <code>Christianity in the United States Christ is the fifth-largest denomination, the largest Pentecostal church, and the largest traditionally African-American denomination in the nation. Among Eastern Christian denominations, there are several Eastern Orthodox and Oriental Orthodox churches, with just below 1 million adherents in the US, or 0.4% of the total population. Christianity was introduced to the Americas as it was first colonized by Europeans beginning in the 16th and 17th centuries. Going forward from its foundation, the United States has been called a Protestant nation by a variety of sources. Immigration further increased Christian numbers. Today most Christian churches in the United States are either</code> | <code>What Will Happen to the Gang Next Year? What Will Happen to the Gang Next Year? "What Will Happen to the Gang Next Year?" is the twenty-second and final episode of the sixth season of the American television comedy series "30 Rock", and the 125th overall episode of the series. It was directed by Michael Engler, and written by Matt Hubbard. The episode originally aired on the National Broadcasting Company (NBC) network in the United States on May 17, 2012. In the episode, Jack (Alec Baldwin) and Avery (Elizabeth Banks) seek to renew their vows; Criss (James Marsden) sets out to show Liz (Tina Fey) he can pay</code> | <code>History of the Jews in the United States Representatives by Rep. Samuel Dickstein (D; New York). This also failed to pass. During the Holocaust, fewer than 30,000 Jews a year reached the United States, and some were turned away due to immigration policies. The U.S. did not change its immigration policies until 1948. Currently, laws requiring teaching of the Holocaust are on the books in five states. The Holocaust had a profound impact on the community in the United States, especially after 1960, as Jews tried to comprehend what had happened, and especially to commemorate and grapple with it when looking to the future. Abraham Joshua Heschel summarized</code> | <code>Public holidays in the United States will have very few customers that day. The labor force in the United States comprises about 62% (as of 2014) of the general population. In the United States, 97% of the private sector businesses determine what days this sector of the population gets paid time off, according to a study by the Society for Human Resource Management. The following holidays are observed by the majority of US businesses with paid time off: This list of holidays is based off the official list of federal holidays by year from the US Government. The holidays however are at the discretion of employers</code> | | <code>Which is the largest species of the turtle family?</code> | <code>Loggerhead sea turtle turtle is debated, but most authors consider it a single polymorphic species. Molecular genetics has confirmed hybridization of the loggerhead sea turtle with the Kemp's ridley sea turtle, hawksbill sea turtle, and green sea turtles. The extent of natural hybridization is not yet determined; however, second-generation hybrids have been reported, suggesting some hybrids are fertile. Although evidence is lacking, modern sea turtles probably descended from a single common ancestor during the Cretaceous period. Like all other sea turtles except the leatherback, loggerheads are members of the ancient family Cheloniidae, and appeared about 40 million years ago. Of the six species</code> | <code>Convention on the Conservation of Migratory Species of Wild Animals take joint action. At May 2018, there were 126 Parties to the Convention. The CMS Family covers a great diversity of migratory species. The Appendices of CMS include many mammals, including land mammals, marine mammals and bats; birds; fish; reptiles and one insect. Among the instruments, AEWA covers 254 species of birds that are ecologically dependent on wetlands for at least part of their annual cycle. EUROBATS covers 52 species of bat, the Memorandum of Understanding on the Conservation of Migratory Sharks seven species of shark, the IOSEA Marine Turtle MOU six species of marine turtle and the Raptors MoU</code> | <code>Razor-backed musk turtle Razor-backed musk turtle The razor-backed musk turtle ("Sternotherus carinatus") is a species of turtle in the family Kinosternidae. The species is native to the southern United States. There are no subspecies that are recognized as being valid. "S. carinatus" is found in the states of Alabama, Arkansas, Louisiana, Mississippi, Oklahoma, and Texas. The razor-backed musk turtle grows to a straight carapace length of about . It has a brown-colored carapace, with black markings at the edges of each scute. The carapace has a distinct, sharp keel down the center of its length, giving the species its common name. The body</code> | <code>African helmeted turtle African helmeted turtle The African helmeted turtle ("Pelomedusa subrufa"), also known commonly as the marsh terrapin, the crocodile turtle, or in the pet trade as the African side-necked turtle, is a species of omnivorous side-necked terrapin in the family Pelomedusidae. The species naturally occurs in fresh and stagnant water bodies throughout much of Sub-Saharan Africa, and in southern Yemen. The marsh terrapin is typically a rather small turtle, with most individuals being less than in straight carapace length, but one has been recorded with a length of . It has a black or brown carapace. The top of the tail</code> | <code>Box turtle Box turtle Box turtles are North American turtles of the genus Terrapene. Although box turtles are superficially similar to tortoises in terrestrial habits and overall appearance, they are actually members of the American pond turtle family (Emydidae). The twelve taxa which are distinguished in the genus are distributed over four species. They are largely characterized by having a domed shell, which is hinged at the bottom, allowing the animal to close its shell tightly to escape predators. The genus name "Terrapene" was coined by Merrem in 1820 as a genus separate from "Emys" for those species which had a sternum</code> | <code>Vallarta mud turtle Vallarta mud turtle The Vallarta mud turtle ("Kinosternon vogti") is a recently identified species of mud turtle in the family Kinosternidae. While formerly considered conspecific with the Jalisco mud turtle, further studies indicated that it was a separate species. It can be identified by a combination of the number of plastron and carapace scutes, body size, and the distinctive yellow rostral shield in males. It is endemic to Mexican state of Jalisco. It is only known from a few human-created or human-affected habitats (such as small streams and ponds) found around Puerto Vallarta. It is one of only 3 species</code> | | <code>How many gallons of beer are in an English barrel?</code> | <code>Low-alcohol beer Prohibition in the United States. Near beer could not legally be labeled as "beer" and was officially classified as a "cereal beverage". The public, however, almost universally called it "near beer". The most popular "near beer" was Bevo, brewed by the Anheuser-Busch company. The Pabst company brewed "Pablo", Miller brewed "Vivo", and Schlitz brewed "Famo". Many local and regional breweries stayed in business by marketing their own near-beers. By 1921 production of near beer had reached over 300 million US gallons (1 billion L) a year (36 L/s). A popular illegal practice was to add alcohol to near beer. The</code> | <code>Keg terms "half-barrel" and "quarter-barrel" are derived from the U.S. beer barrel, legally defined as being equal to 31 U.S. gallons (this is not the same volume as some other units also known as "barrels"). A 15.5 U.S. gallon keg is also equal to: However, beer kegs can come in many sizes: In European countries the most common keg size is 50 liters. This includes the UK, which uses a non-metric standard keg of 11 imperial gallons, which is coincidentally equal to . The German DIN 6647-1 and DIN 6647-2 have also defined kegs in the sizes of 30 and 20</code> | <code>Beer in Chile craft beers. They are generally low or very low volume producers. In Chile there are more than 150 craft beer producers distributed along the 15 Chilean Regions. The list below includes: Beer in Chile The primary beer brewed and consumed in Chile is pale lager, though the country also has a tradition of brewing corn beer, known as chicha. Chile’s beer history has a strong German influence – some of the bigger beer producers are from the country’s southern lake district, a region populated by a great number of German immigrants during the 19th century. Chile also produces English ale-style</code> | <code>Barrel variation. In modern times, produce barrels for all dry goods, excepting cranberries, contain 7,056 cubic inches, about 115.627 L. Barrel A barrel, cask, or tun is a hollow cylindrical container, traditionally made of wooden staves bound by wooden or metal hoops. Traditionally, the barrel was a standard size of measure referring to a set capacity or weight of a given commodity. For example, in the UK a barrel of beer refers to a quantity of . Wine was shipped in barrels of . Modern wooden barrels for wine-making are either made of French common oak ("Quercus robur") and white oak</code> | <code>The Rare Barrel The Rare Barrel The Rare Barrel is a brewery and brewpub in Berkeley, California, United States, that exclusively produces sour beers. Founders Jay Goodwin and Alex Wallash met while attending UCSB. They started home-brewing in their apartment and decided that they would one day start a brewery together. Goodwin started working at The Bruery, where he worked his way from a production assistant to brewer, eventually becoming the head of their barrel aging program. The Rare Barrel brewed its first batch of beer in February 2013, and opened its tasting room on December 27, 2013. The Rare Barrel was named</code> | <code>Barrel (unit) Barrel (unit) A barrel is one of several units of volume applied in various contexts; there are dry barrels, fluid barrels (such as the UK beer barrel and US beer barrel), oil barrels and so on. For historical reasons the volumes of some barrel units are roughly double the volumes of others; volumes in common usage range from about . In many connections the term "drum" is used almost interchangeably with "barrel". Since medieval times the term barrel as a unit of measure has had various meanings throughout Europe, ranging from about 100 litres to 1000 litres. The name was</code> | * Loss: [<code>CachedGISTEmbedLoss</code>](https://sbert.net/docs/package_reference/sentence_transformer/losses.html#cachedgistembedloss) with these parameters: ```json {'guide': SentenceTransformer( (0): Transformer({'max_seq_length': 1024, 'do_lower_case': False}) with Transformer model: XLMRobertaModel (1): Pooling({'word_embedding_dimension': 1024, 'pooling_mode_cls_token': True, 'pooling_mode_mean_tokens': False, 'pooling_mode_max_tokens': False, 'pooling_mode_mean_sqrt_len_tokens': False, 'pooling_mode_weightedmean_tokens': False, 'pooling_mode_lasttoken': False, 'include_prompt': True}) (2): Normalize() ), 'temperature': 0.01} ``` ### Training Hyperparameters #### Non-Default Hyperparameters - `per_device_train_batch_size`: 1024 - `learning_rate`: 3e-05 - `weight_decay`: 0.01 - `warmup_ratio`: 0.05 - `bf16`: True - `batch_sampler`: no_duplicates #### All Hyperparameters <details><summary>Click to expand</summary> - `overwrite_output_dir`: False - `do_predict`: False - `eval_strategy`: no - `prediction_loss_only`: True - `per_device_train_batch_size`: 1024 - `per_device_eval_batch_size`: 8 - `per_gpu_train_batch_size`: None - `per_gpu_eval_batch_size`: None - `gradient_accumulation_steps`: 1 - `eval_accumulation_steps`: None - `torch_empty_cache_steps`: None - `learning_rate`: 3e-05 - `weight_decay`: 0.01 - `adam_beta1`: 0.9 - `adam_beta2`: 0.999 - `adam_epsilon`: 1e-08 - `max_grad_norm`: 1.0 - `num_train_epochs`: 3 - `max_steps`: -1 - `lr_scheduler_type`: linear - `lr_scheduler_kwargs`: {} - `warmup_ratio`: 0.05 - `warmup_steps`: 0 - `log_level`: passive - `log_level_replica`: warning - `log_on_each_node`: True - `logging_nan_inf_filter`: True - `save_safetensors`: True - `save_on_each_node`: False - `save_only_model`: False - `restore_callback_states_from_checkpoint`: False - `no_cuda`: False - `use_cpu`: False - `use_mps_device`: False - `seed`: 42 - `data_seed`: None - `jit_mode_eval`: False - `use_ipex`: False - `bf16`: True - `fp16`: False - `fp16_opt_level`: O1 - `half_precision_backend`: auto - `bf16_full_eval`: False - `fp16_full_eval`: False - `tf32`: None - `local_rank`: 0 - `ddp_backend`: None - `tpu_num_cores`: None - `tpu_metrics_debug`: False - `debug`: [] - `dataloader_drop_last`: True - `dataloader_num_workers`: 0 - `dataloader_prefetch_factor`: None - `past_index`: -1 - `disable_tqdm`: False - `remove_unused_columns`: True - `label_names`: None - `load_best_model_at_end`: False - `ignore_data_skip`: False - `fsdp`: [] - `fsdp_min_num_params`: 0 - `fsdp_config`: {'min_num_params': 0, 'xla': False, 'xla_fsdp_v2': False, 'xla_fsdp_grad_ckpt': False} - `fsdp_transformer_layer_cls_to_wrap`: None - `accelerator_config`: {'split_batches': False, 'dispatch_batches': None, 'even_batches': True, 'use_seedable_sampler': True, 'non_blocking': False, 'gradient_accumulation_kwargs': None} - `deepspeed`: None - `label_smoothing_factor`: 0.0 - `optim`: adamw_torch - `optim_args`: None - `adafactor`: False - `group_by_length`: False - `length_column_name`: length - `ddp_find_unused_parameters`: None - `ddp_bucket_cap_mb`: None - `ddp_broadcast_buffers`: False - `dataloader_pin_memory`: True - `dataloader_persistent_workers`: False - `skip_memory_metrics`: True - `use_legacy_prediction_loop`: False - `push_to_hub`: False - `resume_from_checkpoint`: None - `hub_model_id`: None - `hub_strategy`: every_save - `hub_private_repo`: None - `hub_always_push`: False - `gradient_checkpointing`: False - `gradient_checkpointing_kwargs`: None - `include_inputs_for_metrics`: False - `include_for_metrics`: [] - `eval_do_concat_batches`: True - `fp16_backend`: auto - `push_to_hub_model_id`: None - `push_to_hub_organization`: None - `mp_parameters`: - `auto_find_batch_size`: False - `full_determinism`: False - `torchdynamo`: None - `ray_scope`: last - `ddp_timeout`: 1800 - `torch_compile`: False - `torch_compile_backend`: None - `torch_compile_mode`: None - `dispatch_batches`: None - `split_batches`: None - `include_tokens_per_second`: False - `include_num_input_tokens_seen`: False - `neftune_noise_alpha`: None - `optim_target_modules`: None - `batch_eval_metrics`: False - `eval_on_start`: False - `use_liger_kernel`: False - `eval_use_gather_object`: False - `average_tokens_across_devices`: False - `prompts`: None - `batch_sampler`: no_duplicates - `multi_dataset_batch_sampler`: proportional </details> ### Training Logs <details><summary>Click to expand</summary> | Epoch | Step | Training Loss | |:------:|:----:|:-------------:| | 0.0036 | 1 | 1.0283 | | 0.0072 | 2 | 1.0155 | | 0.0108 | 3 | 0.9858 | | 0.0144 | 4 | 0.9519 | | 0.0181 | 5 | 0.9434 | | 0.0217 | 6 | 0.898 | | 0.0253 | 7 | 0.8798 | | 0.0289 | 8 | 0.7976 | | 0.0325 | 9 | 0.7797 | | 0.0361 | 10 | 0.7464 | | 0.0397 | 11 | 0.743 | | 0.0433 | 12 | 0.716 | | 0.0469 | 13 | 0.7076 | | 0.0505 | 14 | 0.666 | | 0.0542 | 15 | 0.631 | | 0.0578 | 16 | 0.5905 | | 0.0614 | 17 | 0.6537 | | 0.0650 | 18 | 0.5755 | | 0.0686 | 19 | 0.5422 | | 0.0722 | 20 | 0.5393 | | 0.0758 | 21 | 0.5741 | | 0.0794 | 22 | 0.498 | | 0.0830 | 23 | 0.5522 | | 0.0866 | 24 | 0.5592 | | 0.0903 | 25 | 0.4797 | | 0.0939 | 26 | 0.4684 | | 0.0975 | 27 | 0.5207 | | 0.1011 | 28 | 0.4692 | | 0.1047 | 29 | 0.4459 | | 0.1083 | 30 | 0.4439 | | 0.1119 | 31 | 0.4656 | | 0.1155 | 32 | 0.4737 | | 0.1191 | 33 | 0.4391 | | 0.1227 | 34 | 0.4386 | | 0.1264 | 35 | 0.4107 | | 0.1300 | 36 | 0.4513 | | 0.1336 | 37 | 0.3789 | | 0.1372 | 38 | 0.4103 | | 0.1408 | 39 | 0.3929 | | 0.1444 | 40 | 0.4226 | | 0.1480 | 41 | 0.391 | | 0.1516 | 42 | 0.3674 | | 0.1552 | 43 | 0.3607 | | 0.1588 | 44 | 0.3738 | | 0.1625 | 45 | 0.3842 | | 0.1661 | 46 | 0.3498 | | 0.1697 | 47 | 0.3586 | | 0.1733 | 48 | 0.3538 | | 0.1769 | 49 | 0.3572 | | 0.1805 | 50 | 0.3547 | | 0.1841 | 51 | 0.3179 | | 0.1877 | 52 | 0.3436 | | 0.1913 | 53 | 0.3502 | | 0.1949 | 54 | 0.3381 | | 0.1986 | 55 | 0.3547 | | 0.2022 | 56 | 0.3362 | | 0.2058 | 57 | 0.3407 | | 0.2094 | 58 | 0.31 | | 0.2130 | 59 | 0.3039 | | 0.2166 | 60 | 0.3362 | | 0.2202 | 61 | 0.2948 | | 0.2238 | 62 | 0.3429 | | 0.2274 | 63 | 0.3096 | | 0.2310 | 64 | 0.35 | | 0.2347 | 65 | 0.2997 | | 0.2383 | 66 | 0.3258 | | 0.2419 | 67 | 0.3376 | | 0.2455 | 68 | 0.3213 | | 0.2491 | 69 | 0.3185 | | 0.2527 | 70 | 0.3282 | | 0.2563 | 71 | 0.2988 | | 0.2599 | 72 | 0.33 | | 0.2635 | 73 | 0.3066 | | 0.2671 | 74 | 0.3303 | | 0.2708 | 75 | 0.3067 | | 0.2744 | 76 | 0.2996 | | 0.2780 | 77 | 0.3063 | | 0.2816 | 78 | 0.3235 | | 0.2852 | 79 | 0.2902 | | 0.2888 | 80 | 0.302 | | 0.2924 | 81 | 0.3223 | | 0.2960 | 82 | 0.297 | | 0.2996 | 83 | 0.2936 | | 0.3032 | 84 | 0.3279 | | 0.3069 | 85 | 0.2973 | | 0.3105 | 86 | 0.2881 | | 0.3141 | 87 | 0.3014 | | 0.3177 | 88 | 0.2986 | | 0.3213 | 89 | 0.3057 | | 0.3249 | 90 | 0.2887 | | 0.3285 | 91 | 0.2765 | | 0.3321 | 92 | 0.2818 | | 0.3357 | 93 | 0.2904 | | 0.3394 | 94 | 0.267 | | 0.3430 | 95 | 0.2948 | | 0.3466 | 96 | 0.2766 | | 0.3502 | 97 | 0.2782 | | 0.3538 | 98 | 0.3082 | | 0.3574 | 99 | 0.2697 | | 0.3610 | 100 | 0.3006 | | 0.3646 | 101 | 0.2986 | | 0.3682 | 102 | 0.2789 | | 0.3718 | 103 | 0.2756 | | 0.3755 | 104 | 0.2884 | | 0.3791 | 105 | 0.273 | | 0.3827 | 106 | 0.2687 | | 0.3863 | 107 | 0.2808 | | 0.3899 | 108 | 0.2763 | | 0.3935 | 109 | 0.2738 | | 0.3971 | 110 | 0.2642 | | 0.4007 | 111 | 0.2612 | | 0.4043 | 112 | 0.2859 | | 0.4079 | 113 | 0.2558 | | 0.4116 | 114 | 0.2565 | | 0.4152 | 115 | 0.2747 | | 0.4188 | 116 | 0.2684 | | 0.4224 | 117 | 0.2643 | | 0.4260 | 118 | 0.241 | | 0.4296 | 119 | 0.2563 | | 0.4332 | 120 | 0.2754 | | 0.4368 | 121 | 0.2503 | | 0.4404 | 122 | 0.2544 | | 0.4440 | 123 | 0.2729 | | 0.4477 | 124 | 0.2589 | | 0.4513 | 125 | 0.2626 | | 0.4549 | 126 | 0.2693 | | 0.4585 | 127 | 0.2687 | | 0.4621 | 128 | 0.2903 | | 0.4657 | 129 | 0.2663 | | 0.4693 | 130 | 0.2604 | | 0.4729 | 131 | 0.2601 | | 0.4765 | 132 | 0.2649 | | 0.4801 | 133 | 0.2597 | | 0.4838 | 134 | 0.2608 | | 0.4874 | 135 | 0.245 | | 0.4910 | 136 | 0.2587 | | 0.4946 | 137 | 0.2618 | | 0.4982 | 138 | 0.2599 | | 0.5018 | 139 | 0.265 | | 0.5054 | 140 | 0.2427 | | 0.5090 | 141 | 0.2448 | | 0.5126 | 142 | 0.2608 | | 0.5162 | 143 | 0.2188 | | 0.5199 | 144 | 0.2471 | | 0.5235 | 145 | 0.2604 | | 0.5271 | 146 | 0.2571 | | 0.5307 | 147 | 0.2684 | | 0.5343 | 148 | 0.2319 | | 0.5379 | 149 | 0.2572 | | 0.5415 | 150 | 0.2243 | | 0.5451 | 151 | 0.2562 | | 0.5487 | 152 | 0.2457 | | 0.5523 | 153 | 0.255 | | 0.5560 | 154 | 0.2664 | | 0.5596 | 155 | 0.24 | | 0.5632 | 156 | 0.2612 | | 0.5668 | 157 | 0.243 | | 0.5704 | 158 | 0.2345 | | 0.5740 | 159 | 0.2359 | | 0.5776 | 160 | 0.2384 | | 0.5812 | 161 | 0.2541 | | 0.5848 | 162 | 0.2496 | | 0.5884 | 163 | 0.2429 | | 0.5921 | 164 | 0.2411 | | 0.5957 | 165 | 0.2261 | | 0.5993 | 166 | 0.2164 | | 0.6029 | 167 | 0.2251 | | 0.6065 | 168 | 0.2417 | | 0.6101 | 169 | 0.2494 | | 0.6137 | 170 | 0.2359 | | 0.6173 | 171 | 0.2489 | | 0.6209 | 172 | 0.2261 | | 0.6245 | 173 | 0.2367 | | 0.6282 | 174 | 0.2355 | | 0.6318 | 175 | 0.2423 | | 0.6354 | 176 | 0.2454 | | 0.6390 | 177 | 0.2438 | | 0.6426 | 178 | 0.2415 | | 0.6462 | 179 | 0.2237 | | 0.6498 | 180 | 0.2419 | | 0.6534 | 181 | 0.2373 | | 0.6570 | 182 | 0.2659 | | 0.6606 | 183 | 0.2201 | | 0.6643 | 184 | 0.2342 | | 0.6679 | 185 | 0.2149 | | 0.6715 | 186 | 0.2241 | | 0.6751 | 187 | 0.2443 | | 0.6787 | 188 | 0.2489 | | 0.6823 | 189 | 0.2354 | | 0.6859 | 190 | 0.2483 | | 0.6895 | 191 | 0.2193 | | 0.6931 | 192 | 0.229 | | 0.6968 | 193 | 0.2335 | | 0.7004 | 194 | 0.2484 | | 0.7040 | 195 | 0.2317 | | 0.7076 | 196 | 0.2203 | | 0.7112 | 197 | 0.2329 | | 0.7148 | 198 | 0.2084 | | 0.7184 | 199 | 0.2341 | | 0.7220 | 200 | 0.2369 | | 0.7256 | 201 | 0.2364 | | 0.7292 | 202 | 0.2276 | | 0.7329 | 203 | 0.215 | | 0.7365 | 204 | 0.2486 | | 0.7401 | 205 | 0.2237 | | 0.7437 | 206 | 0.218 | | 0.7473 | 207 | 0.2444 | | 0.7509 | 208 | 0.2276 | | 0.7545 | 209 | 0.2127 | | 0.7581 | 210 | 0.2283 | | 0.7617 | 211 | 0.2234 | | 0.7653 | 212 | 0.207 | | 0.7690 | 213 | 0.24 | | 0.7726 | 214 | 0.2317 | | 0.7762 | 215 | 0.2056 | | 0.7798 | 216 | 0.2149 | | 0.7834 | 217 | 0.2211 | | 0.7870 | 218 | 0.2232 | | 0.7906 | 219 | 0.2222 | | 0.7942 | 220 | 0.2481 | | 0.7978 | 221 | 0.227 | | 0.8014 | 222 | 0.2305 | | 0.8051 | 223 | 0.2091 | | 0.8087 | 224 | 0.2278 | | 0.8123 | 225 | 0.2123 | | 0.8159 | 226 | 0.2233 | | 0.8195 | 227 | 0.2365 | | 0.8231 | 228 | 0.2165 | | 0.8267 | 229 | 0.2192 | | 0.8303 | 230 | 0.2145 | | 0.8339 | 231 | 0.2382 | | 0.8375 | 232 | 0.2232 | | 0.8412 | 233 | 0.2273 | | 0.8448 | 234 | 0.2296 | | 0.8484 | 235 | 0.2229 | | 0.8520 | 236 | 0.2213 | | 0.8556 | 237 | 0.2343 | | 0.8592 | 238 | 0.2208 | | 0.8628 | 239 | 0.2315 | | 0.8664 | 240 | 0.2137 | | 0.8700 | 241 | 0.2201 | | 0.8736 | 242 | 0.2185 | | 0.8773 | 243 | 0.2337 | | 0.8809 | 244 | 0.2153 | | 0.8845 | 245 | 0.2369 | | 0.8881 | 246 | 0.2216 | | 0.8917 | 247 | 0.2338 | | 0.8953 | 248 | 0.2241 | | 0.8989 | 249 | 0.213 | | 0.9025 | 250 | 0.2245 | | 0.9061 | 251 | 0.2074 | | 0.9097 | 252 | 0.2283 | | 0.9134 | 253 | 0.2003 | | 0.9170 | 254 | 0.2099 | | 0.9206 | 255 | 0.2288 | | 0.9242 | 256 | 0.2168 | | 0.9278 | 257 | 0.215 | | 0.9314 | 258 | 0.2146 | | 0.9350 | 259 | 0.2126 | | 0.9386 | 260 | 0.2178 | | 0.9422 | 261 | 0.2065 | | 0.9458 | 262 | 0.2327 | | 0.9495 | 263 | 0.2116 | | 0.9531 | 264 | 0.2324 | | 0.9567 | 265 | 0.2235 | | 0.9603 | 266 | 0.2189 | | 0.9639 | 267 | 0.2175 | | 0.9675 | 268 | 0.2171 | | 0.9711 | 269 | 0.1925 | | 0.9747 | 270 | 0.225 | | 0.9783 | 271 | 0.2149 | | 0.9819 | 272 | 0.204 | | 0.9856 | 273 | 0.2004 | | 0.9892 | 274 | 0.2055 | | 0.9928 | 275 | 0.2045 | | 0.9964 | 276 | 0.2186 | | 1.0 | 277 | 0.2215 | | 1.0036 | 278 | 0.1545 | | 1.0072 | 279 | 0.169 | | 1.0108 | 280 | 0.152 | | 1.0144 | 281 | 0.1597 | | 1.0181 | 282 | 0.1626 | | 1.0217 | 283 | 0.1692 | | 1.0253 | 284 | 0.1639 | | 1.0289 | 285 | 0.1638 | | 1.0325 | 286 | 0.1507 | | 1.0361 | 287 | 0.1594 | | 1.0397 | 288 | 0.1621 | | 1.0433 | 289 | 0.1565 | | 1.0469 | 290 | 0.1549 | | 1.0505 | 291 | 0.1731 | | 1.0542 | 292 | 0.152 | | 1.0578 | 293 | 0.1586 | | 1.0614 | 294 | 0.1593 | | 1.0650 | 295 | 0.1406 | | 1.0686 | 296 | 0.1524 | | 1.0722 | 297 | 0.1474 | | 1.0758 | 298 | 0.158 | | 1.0794 | 299 | 0.1743 | | 1.0830 | 300 | 0.1485 | | 1.0866 | 301 | 0.1648 | | 1.0903 | 302 | 0.1337 | | 1.0939 | 303 | 0.1554 | | 1.0975 | 304 | 0.1434 | | 1.1011 | 305 | 0.1642 | | 1.1047 | 306 | 0.159 | | 1.1083 | 307 | 0.1658 | | 1.1119 | 308 | 0.1554 | | 1.1155 | 309 | 0.1425 | | 1.1191 | 310 | 0.1432 | | 1.1227 | 311 | 0.1517 | | 1.1264 | 312 | 0.148 | | 1.1300 | 313 | 0.1636 | | 1.1336 | 314 | 0.1735 | | 1.1372 | 315 | 0.151 | | 1.1408 | 316 | 0.1423 | | 1.1444 | 317 | 0.1501 | | 1.1480 | 318 | 0.1537 | | 1.1516 | 319 | 0.1554 | | 1.1552 | 320 | 0.1553 | | 1.1588 | 321 | 0.149 | | 1.1625 | 322 | 0.1605 | | 1.1661 | 323 | 0.1551 | | 1.1697 | 324 | 0.1555 | | 1.1733 | 325 | 0.1443 | | 1.1769 | 326 | 0.1533 | | 1.1805 | 327 | 0.1658 | | 1.1841 | 328 | 0.15 | | 1.1877 | 329 | 0.1626 | | 1.1913 | 330 | 0.172 | | 1.1949 | 331 | 0.1542 | | 1.1986 | 332 | 0.166 | | 1.2022 | 333 | 0.1513 | | 1.2058 | 334 | 0.1612 | | 1.2094 | 335 | 0.1521 | | 1.2130 | 336 | 0.1552 | | 1.2166 | 337 | 0.1503 | | 1.2202 | 338 | 0.1613 | | 1.2238 | 339 | 0.1563 | | 1.2274 | 340 | 0.1429 | | 1.2310 | 341 | 0.1587 | | 1.2347 | 342 | 0.1477 | | 1.2383 | 343 | 0.1561 | | 1.2419 | 344 | 0.1418 | | 1.2455 | 345 | 0.1495 | | 1.2491 | 346 | 0.1533 | | 1.2527 | 347 | 0.1521 | | 1.2563 | 348 | 0.1422 | | 1.2599 | 349 | 0.1446 | | 1.2635 | 350 | 0.146 | | 1.2671 | 351 | 0.1473 | | 1.2708 | 352 | 0.1566 | | 1.2744 | 353 | 0.1411 | | 1.2780 | 354 | 0.1502 | | 1.2816 | 355 | 0.1383 | | 1.2852 | 356 | 0.1622 | | 1.2888 | 357 | 0.1391 | | 1.2924 | 358 | 0.1455 | | 1.2960 | 359 | 0.1541 | | 1.2996 | 360 | 0.1476 | | 1.3032 | 361 | 0.1662 | | 1.3069 | 362 | 0.1476 | | 1.3105 | 363 | 0.1452 | | 1.3141 | 364 | 0.1372 | | 1.3177 | 365 | 0.1542 | | 1.3213 | 366 | 0.1531 | | 1.3249 | 367 | 0.1623 | | 1.3285 | 368 | 0.1544 | | 1.3321 | 369 | 0.1625 | | 1.3357 | 370 | 0.1459 | | 1.3394 | 371 | 0.1474 | | 1.3430 | 372 | 0.1499 | | 1.3466 | 373 | 0.1495 | | 1.3502 | 374 | 0.1361 | | 1.3538 | 375 | 0.1444 | | 1.3574 | 376 | 0.1495 | | 1.3610 | 377 | 0.1583 | | 1.3646 | 378 | 0.1642 | | 1.3682 | 379 | 0.1646 | | 1.3718 | 380 | 0.1595 | | 1.3755 | 381 | 0.149 | | 1.3791 | 382 | 0.1448 | | 1.3827 | 383 | 0.1603 | | 1.3863 | 384 | 0.1269 | | 1.3899 | 385 | 0.1491 | | 1.3935 | 386 | 0.1367 | | 1.3971 | 387 | 0.1501 | | 1.4007 | 388 | 0.1414 | | 1.4043 | 389 | 0.156 | | 1.4079 | 390 | 0.1428 | | 1.4116 | 391 | 0.1559 | | 1.4152 | 392 | 0.1452 | | 1.4188 | 393 | 0.1547 | | 1.4224 | 394 | 0.1432 | | 1.4260 | 395 | 0.1648 | | 1.4296 | 396 | 0.166 | | 1.4332 | 397 | 0.1485 | | 1.4368 | 398 | 0.1494 | | 1.4404 | 399 | 0.1635 | | 1.4440 | 400 | 0.1498 | | 1.4477 | 401 | 0.1509 | | 1.4513 | 402 | 0.1431 | | 1.4549 | 403 | 0.1547 | | 1.4585 | 404 | 0.1576 | | 1.4621 | 405 | 0.1426 | | 1.4657 | 406 | 0.132 | | 1.4693 | 407 | 0.1511 | | 1.4729 | 408 | 0.1551 | | 1.4765 | 409 | 0.16 | | 1.4801 | 410 | 0.1507 | | 1.4838 | 411 | 0.1591 | | 1.4874 | 412 | 0.1536 | | 1.4910 | 413 | 0.1507 | | 1.4946 | 414 | 0.1564 | | 1.4982 | 415 | 0.153 | | 1.5018 | 416 | 0.1404 | | 1.5054 | 417 | 0.1627 | | 1.5090 | 418 | 0.1432 | | 1.5126 | 419 | 0.1456 | | 1.5162 | 420 | 0.1369 | | 1.5199 | 421 | 0.1554 | | 1.5235 | 422 | 0.1412 | | 1.5271 | 423 | 0.1547 | | 1.5307 | 424 | 0.1555 | | 1.5343 | 425 | 0.1575 | | 1.5379 | 426 | 0.1595 | | 1.5415 | 427 | 0.1464 | | 1.5451 | 428 | 0.1738 | | 1.5487 | 429 | 0.1692 | | 1.5523 | 430 | 0.1566 | | 1.5560 | 431 | 0.1452 | | 1.5596 | 432 | 0.1433 | | 1.5632 | 433 | 0.1584 | | 1.5668 | 434 | 0.1579 | | 1.5704 | 435 | 0.157 | | 1.5740 | 436 | 0.1533 | | 1.5776 | 437 | 0.148 | | 1.5812 | 438 | 0.1381 | | 1.5848 | 439 | 0.1605 | | 1.5884 | 440 | 0.163 | | 1.5921 | 441 | 0.1492 | | 1.5957 | 442 | 0.1601 | | 1.5993 | 443 | 0.1456 | | 1.6029 | 444 | 0.1439 | | 1.6065 | 445 | 0.1553 | | 1.6101 | 446 | 0.1371 | | 1.6137 | 447 | 0.1382 | | 1.6173 | 448 | 0.1458 | | 1.6209 | 449 | 0.14 | | 1.6245 | 450 | 0.1463 | | 1.6282 | 451 | 0.1433 | | 1.6318 | 452 | 0.1472 | | 1.6354 | 453 | 0.1481 | | 1.6390 | 454 | 0.1408 | | 1.6426 | 455 | 0.1525 | | 1.6462 | 456 | 0.1223 | | 1.6498 | 457 | 0.1452 | | 1.6534 | 458 | 0.159 | | 1.6570 | 459 | 0.1389 | | 1.6606 | 460 | 0.1479 | | 1.6643 | 461 | 0.1451 | | 1.6679 | 462 | 0.1651 | | 1.6715 | 463 | 0.1336 | | 1.6751 | 464 | 0.1496 | | 1.6787 | 465 | 0.1384 | | 1.6823 | 466 | 0.143 | | 1.6859 | 467 | 0.1423 | | 1.6895 | 468 | 0.1403 | | 1.6931 | 469 | 0.1577 | | 1.6968 | 470 | 0.1511 | | 1.7004 | 471 | 0.1429 | | 1.7040 | 472 | 0.1445 | | 1.7076 | 473 | 0.1431 | | 1.7112 | 474 | 0.1326 | | 1.7148 | 475 | 0.1554 | | 1.7184 | 476 | 0.1406 | | 1.7220 | 477 | 0.1479 | | 1.7256 | 478 | 0.1521 | | 1.7292 | 479 | 0.1475 | | 1.7329 | 480 | 0.1584 | | 1.7365 | 481 | 0.1393 | | 1.7401 | 482 | 0.1291 | | 1.7437 | 483 | 0.1373 | | 1.7473 | 484 | 0.1555 | | 1.7509 | 485 | 0.1473 | | 1.7545 | 486 | 0.1654 | | 1.7581 | 487 | 0.1568 | | 1.7617 | 488 | 0.1557 | | 1.7653 | 489 | 0.1531 | | 1.7690 | 490 | 0.1385 | | 1.7726 | 491 | 0.1381 | | 1.7762 | 492 | 0.1375 | | 1.7798 | 493 | 0.1472 | | 1.7834 | 494 | 0.1581 | | 1.7870 | 495 | 0.1448 | | 1.7906 | 496 | 0.1443 | | 1.7942 | 497 | 0.1422 | | 1.7978 | 498 | 0.1295 | | 1.8014 | 499 | 0.1463 | | 1.8051 | 500 | 0.1346 | | 1.8087 | 501 | 0.1387 | | 1.8123 | 502 | 0.1463 | | 1.8159 | 503 | 0.1439 | | 1.8195 | 504 | 0.1404 | | 1.8231 | 505 | 0.1433 | | 1.8267 | 506 | 0.136 | | 1.8303 | 507 | 0.14 | | 1.8339 | 508 | 0.1355 | | 1.8375 | 509 | 0.1446 | | 1.8412 | 510 | 0.1564 | | 1.8448 | 511 | 0.1413 | | 1.8484 | 512 | 0.1451 | | 1.8520 | 513 | 0.1453 | | 1.8556 | 514 | 0.1484 | | 1.8592 | 515 | 0.1403 | | 1.8628 | 516 | 0.1568 | | 1.8664 | 517 | 0.1566 | | 1.8700 | 518 | 0.1318 | | 1.8736 | 519 | 0.1483 | | 1.8773 | 520 | 0.1339 | | 1.8809 | 521 | 0.1423 | | 1.8845 | 522 | 0.1349 | | 1.8881 | 523 | 0.1302 | | 1.8917 | 524 | 0.1341 | | 1.8953 | 525 | 0.1456 | | 1.8989 | 526 | 0.1334 | | 1.9025 | 527 | 0.1382 | | 1.9061 | 528 | 0.1462 | | 1.9097 | 529 | 0.1315 | | 1.9134 | 530 | 0.1606 | | 1.9170 | 531 | 0.1308 | | 1.9206 | 532 | 0.1319 | | 1.9242 | 533 | 0.1407 | | 1.9278 | 534 | 0.1385 | | 1.9314 | 535 | 0.1471 | | 1.9350 | 536 | 0.1621 | | 1.9386 | 537 | 0.1436 | | 1.9422 | 538 | 0.151 | | 1.9458 | 539 | 0.1423 | | 1.9495 | 540 | 0.1411 | | 1.9531 | 541 | 0.1535 | | 1.9567 | 542 | 0.143 | | 1.9603 | 543 | 0.149 | | 1.9639 | 544 | 0.1384 | | 1.9675 | 545 | 0.1479 | | 1.9711 | 546 | 0.1452 | | 1.9747 | 547 | 0.1372 | | 1.9783 | 548 | 0.1418 | | 1.9819 | 549 | 0.1443 | | 1.9856 | 550 | 0.1344 | | 1.9892 | 551 | 0.1278 | | 1.9928 | 552 | 0.1447 | | 1.9964 | 553 | 0.1366 | | 2.0 | 554 | 0.141 | | 2.0036 | 555 | 0.1161 | | 2.0072 | 556 | 0.1099 | | 2.0108 | 557 | 0.126 | | 2.0144 | 558 | 0.1163 | | 2.0181 | 559 | 0.1234 | | 2.0217 | 560 | 0.1171 | | 2.0253 | 561 | 0.1073 | | 2.0289 | 562 | 0.1126 | | 2.0325 | 563 | 0.1175 | | 2.0361 | 564 | 0.1086 | | 2.0397 | 565 | 0.1038 | | 2.0433 | 566 | 0.1121 | | 2.0469 | 567 | 0.1154 | | 2.0505 | 568 | 0.0973 | | 2.0542 | 569 | 0.1208 | | 2.0578 | 570 | 0.1064 | | 2.0614 | 571 | 0.1159 | | 2.0650 | 572 | 0.1093 | | 2.0686 | 573 | 0.113 | | 2.0722 | 574 | 0.1033 | | 2.0758 | 575 | 0.1152 | | 2.0794 | 576 | 0.1029 | | 2.0830 | 577 | 0.1204 | | 2.0866 | 578 | 0.1079 | | 2.0903 | 579 | 0.1288 | | 2.0939 | 580 | 0.0998 | | 2.0975 | 581 | 0.1058 | | 2.1011 | 582 | 0.1235 | | 2.1047 | 583 | 0.1059 | | 2.1083 | 584 | 0.0998 | | 2.1119 | 585 | 0.1142 | | 2.1155 | 586 | 0.1082 | | 2.1191 | 587 | 0.0973 | | 2.1227 | 588 | 0.1017 | | 2.1264 | 589 | 0.1045 | | 2.1300 | 590 | 0.123 | | 2.1336 | 591 | 0.1065 | | 2.1372 | 592 | 0.1135 | | 2.1408 | 593 | 0.1027 | | 2.1444 | 594 | 0.1166 | | 2.1480 | 595 | 0.1082 | | 2.1516 | 596 | 0.1113 | | 2.1552 | 597 | 0.1108 | | 2.1588 | 598 | 0.114 | | 2.1625 | 599 | 0.1064 | | 2.1661 | 600 | 0.0955 | | 2.1697 | 601 | 0.113 | | 2.1733 | 602 | 0.1136 | | 2.1769 | 603 | 0.1125 | | 2.1805 | 604 | 0.1146 | | 2.1841 | 605 | 0.1054 | | 2.1877 | 606 | 0.1144 | | 2.1913 | 607 | 0.1038 | | 2.1949 | 608 | 0.1113 | | 2.1986 | 609 | 0.1187 | | 2.2022 | 610 | 0.1166 | | 2.2058 | 611 | 0.1035 | | 2.2094 | 612 | 0.1054 | | 2.2130 | 613 | 0.118 | | 2.2166 | 614 | 0.125 | | 2.2202 | 615 | 0.1142 | | 2.2238 | 616 | 0.1119 | | 2.2274 | 617 | 0.1173 | | 2.2310 | 618 | 0.1024 | | 2.2347 | 619 | 0.105 | | 2.2383 | 620 | 0.1025 | | 2.2419 | 621 | 0.1022 | | 2.2455 | 622 | 0.0995 | | 2.2491 | 623 | 0.1022 | | 2.2527 | 624 | 0.1198 | | 2.2563 | 625 | 0.0995 | | 2.2599 | 626 | 0.1162 | | 2.2635 | 627 | 0.1172 | | 2.2671 | 628 | 0.1037 | | 2.2708 | 629 | 0.1093 | | 2.2744 | 630 | 0.1018 | | 2.2780 | 631 | 0.1168 | | 2.2816 | 632 | 0.1015 | | 2.2852 | 633 | 0.101 | | 2.2888 | 634 | 0.1064 | | 2.2924 | 635 | 0.1185 | | 2.2960 | 636 | 0.1055 | | 2.2996 | 637 | 0.1142 | | 2.3032 | 638 | 0.0966 | | 2.3069 | 639 | 0.1039 | | 2.3105 | 640 | 0.1139 | | 2.3141 | 641 | 0.1181 | | 2.3177 | 642 | 0.1168 | | 2.3213 | 643 | 0.1201 | | 2.3249 | 644 | 0.0984 | | 2.3285 | 645 | 0.1068 | | 2.3321 | 646 | 0.1007 | | 2.3357 | 647 | 0.1179 | | 2.3394 | 648 | 0.1043 | | 2.3430 | 649 | 0.1213 | | 2.3466 | 650 | 0.1027 | | 2.3502 | 651 | 0.1119 | | 2.3538 | 652 | 0.1077 | | 2.3574 | 653 | 0.1061 | | 2.3610 | 654 | 0.1054 | | 2.3646 | 655 | 0.1135 | | 2.3682 | 656 | 0.1136 | | 2.3718 | 657 | 0.1062 | | 2.3755 | 658 | 0.1105 | | 2.3791 | 659 | 0.1157 | | 2.3827 | 660 | 0.1036 | | 2.3863 | 661 | 0.1098 | | 2.3899 | 662 | 0.1195 | | 2.3935 | 663 | 0.1151 | | 2.3971 | 664 | 0.1116 | | 2.4007 | 665 | 0.1086 | | 2.4043 | 666 | 0.1151 | | 2.4079 | 667 | 0.1156 | | 2.4116 | 668 | 0.116 | | 2.4152 | 669 | 0.1055 | | 2.4188 | 670 | 0.1051 | | 2.4224 | 671 | 0.0952 | | 2.4260 | 672 | 0.1012 | | 2.4296 | 673 | 0.1042 | | 2.4332 | 674 | 0.1069 | | 2.4368 | 675 | 0.1148 | | 2.4404 | 676 | 0.0981 | | 2.4440 | 677 | 0.1131 | | 2.4477 | 678 | 0.1026 | | 2.4513 | 679 | 0.1014 | | 2.4549 | 680 | 0.1071 | | 2.4585 | 681 | 0.1171 | | 2.4621 | 682 | 0.1009 | | 2.4657 | 683 | 0.1056 | | 2.4693 | 684 | 0.1107 | | 2.4729 | 685 | 0.1114 | | 2.4765 | 686 | 0.1118 | | 2.4801 | 687 | 0.1166 | | 2.4838 | 688 | 0.1023 | | 2.4874 | 689 | 0.1154 | | 2.4910 | 690 | 0.0968 | | 2.4946 | 691 | 0.1164 | | 2.4982 | 692 | 0.1221 | | 2.5018 | 693 | 0.1131 | | 2.5054 | 694 | 0.1039 | | 2.5090 | 695 | 0.1022 | | 2.5126 | 696 | 0.1052 | | 2.5162 | 697 | 0.1072 | | 2.5199 | 698 | 0.1062 | | 2.5235 | 699 | 0.1035 | | 2.5271 | 700 | 0.107 | | 2.5307 | 701 | 0.1152 | | 2.5343 | 702 | 0.0991 | | 2.5379 | 703 | 0.1139 | | 2.5415 | 704 | 0.1148 | | 2.5451 | 705 | 0.1099 | | 2.5487 | 706 | 0.1064 | | 2.5523 | 707 | 0.1069 | | 2.5560 | 708 | 0.1104 | | 2.5596 | 709 | 0.1157 | | 2.5632 | 710 | 0.1109 | | 2.5668 | 711 | 0.0991 | | 2.5704 | 712 | 0.105 | | 2.5740 | 713 | 0.1104 | | 2.5776 | 714 | 0.1134 | | 2.5812 | 715 | 0.1252 | | 2.5848 | 716 | 0.1205 | | 2.5884 | 717 | 0.112 | | 2.5921 | 718 | 0.1109 | | 2.5957 | 719 | 0.1151 | | 2.5993 | 720 | 0.097 | | 2.6029 | 721 | 0.1018 | | 2.6065 | 722 | 0.1205 | | 2.6101 | 723 | 0.107 | | 2.6137 | 724 | 0.102 | | 2.6173 | 725 | 0.1106 | | 2.6209 | 726 | 0.1068 | | 2.6245 | 727 | 0.1024 | | 2.6282 | 728 | 0.1153 | | 2.6318 | 729 | 0.0984 | | 2.6354 | 730 | 0.1019 | | 2.6390 | 731 | 0.1029 | | 2.6426 | 732 | 0.1147 | | 2.6462 | 733 | 0.1081 | | 2.6498 | 734 | 0.0996 | | 2.6534 | 735 | 0.1133 | | 2.6570 | 736 | 0.1102 | | 2.6606 | 737 | 0.1063 | | 2.6643 | 738 | 0.1119 | | 2.6679 | 739 | 0.1062 | | 2.6715 | 740 | 0.1021 | | 2.6751 | 741 | 0.1058 | | 2.6787 | 742 | 0.1026 | | 2.6823 | 743 | 0.1049 | | 2.6859 | 744 | 0.0894 | | 2.6895 | 745 | 0.1127 | | 2.6931 | 746 | 0.1107 | | 2.6968 | 747 | 0.1134 | | 2.7004 | 748 | 0.103 | | 2.7040 | 749 | 0.1081 | | 2.7076 | 750 | 0.1156 | | 2.7112 | 751 | 0.1092 | | 2.7148 | 752 | 0.1182 | | 2.7184 | 753 | 0.1092 | | 2.7220 | 754 | 0.1077 | | 2.7256 | 755 | 0.1165 | | 2.7292 | 756 | 0.1109 | | 2.7329 | 757 | 0.1061 | | 2.7365 | 758 | 0.1141 | | 2.7401 | 759 | 0.1073 | | 2.7437 | 760 | 0.1074 | | 2.7473 | 761 | 0.1042 | | 2.7509 | 762 | 0.1083 | | 2.7545 | 763 | 0.1011 | | 2.7581 | 764 | 0.1083 | | 2.7617 | 765 | 0.1078 | | 2.7653 | 766 | 0.1333 | | 2.7690 | 767 | 0.107 | | 2.7726 | 768 | 0.1114 | | 2.7762 | 769 | 0.1027 | | 2.7798 | 770 | 0.0976 | | 2.7834 | 771 | 0.1175 | | 2.7870 | 772 | 0.1099 | | 2.7906 | 773 | 0.102 | | 2.7942 | 774 | 0.1083 | | 2.7978 | 775 | 0.0999 | | 2.8014 | 776 | 0.1221 | | 2.8051 | 777 | 0.0996 | | 2.8087 | 778 | 0.0995 | | 2.8123 | 779 | 0.0974 | | 2.8159 | 780 | 0.1098 | | 2.8195 | 781 | 0.1012 | | 2.8231 | 782 | 0.1006 | | 2.8267 | 783 | 0.1047 | | 2.8303 | 784 | 0.1214 | | 2.8339 | 785 | 0.105 | | 2.8375 | 786 | 0.1152 | | 2.8412 | 787 | 0.0897 | | 2.8448 | 788 | 0.1189 | | 2.8484 | 789 | 0.1058 | | 2.8520 | 790 | 0.1051 | | 2.8556 | 791 | 0.1134 | | 2.8592 | 792 | 0.1016 | | 2.8628 | 793 | 0.1015 | | 2.8664 | 794 | 0.1119 | | 2.8700 | 795 | 0.1093 | | 2.8736 | 796 | 0.108 | | 2.8773 | 797 | 0.109 | | 2.8809 | 798 | 0.0922 | | 2.8845 | 799 | 0.1123 | | 2.8881 | 800 | 0.1094 | | 2.8917 | 801 | 0.1075 | | 2.8953 | 802 | 0.1138 | | 2.8989 | 803 | 0.1071 | | 2.9025 | 804 | 0.105 | | 2.9061 | 805 | 0.1153 | | 2.9097 | 806 | 0.1015 | | 2.9134 | 807 | 0.1199 | | 2.9170 | 808 | 0.1025 | | 2.9206 | 809 | 0.1082 | | 2.9242 | 810 | 0.1025 | | 2.9278 | 811 | 0.1073 | | 2.9314 | 812 | 0.1053 | | 2.9350 | 813 | 0.1088 | | 2.9386 | 814 | 0.1064 | | 2.9422 | 815 | 0.119 | | 2.9458 | 816 | 0.1099 | | 2.9495 | 817 | 0.1096 | | 2.9531 | 818 | 0.1101 | | 2.9567 | 819 | 0.086 | | 2.9603 | 820 | 0.1121 | | 2.9639 | 821 | 0.103 | | 2.9675 | 822 | 0.1108 | | 2.9711 | 823 | 0.1132 | | 2.9747 | 824 | 0.1034 | | 2.9783 | 825 | 0.0995 | | 2.9819 | 826 | 0.115 | | 2.9856 | 827 | 0.0966 | | 2.9892 | 828 | 0.1037 | | 2.9928 | 829 | 0.1124 | | 2.9964 | 830 | 0.1071 | | 3.0 | 831 | 0.1017 | </details> ### Framework Versions - Python: 3.10.12 - Sentence Transformers: 3.4.1 - Transformers: 4.49.0 - PyTorch: 2.5.1+cu124 - Accelerate: 1.4.0 - Datasets: 3.3.2 - Tokenizers: 0.21.0 ## Citation ### BibTeX #### Sentence Transformers ```bibtex @inproceedings{reimers-2019-sentence-bert, title = "Sentence-BERT: Sentence Embeddings using Siamese BERT-Networks", author = "Reimers, Nils and Gurevych, Iryna", booktitle = "Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing", month = "11", year = "2019", publisher = "Association for Computational Linguistics", url = "https://arxiv.org/abs/1908.10084", } ``` <!-- ## Glossary *Clearly define terms in order to be accessible across audiences.* --> <!-- ## Model Card Authors *Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.* --> <!-- ## Model Card Contact *Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.* -->
Xenova/sam-vit-base
Xenova
2025-03-07T10:43:03Z
86
0
transformers.js
[ "transformers.js", "onnx", "sam", "mask-generation", "base_model:facebook/sam-vit-base", "base_model:quantized:facebook/sam-vit-base", "region:us" ]
mask-generation
2023-05-06T16:40:37Z
--- base_model: facebook/sam-vit-base library_name: transformers.js --- https://huggingface.co/facebook/sam-vit-base with ONNX weights to be compatible with Transformers.js. ## Usage (Transformers.js) If you haven't already, you can install the [Transformers.js](https://huggingface.co/docs/transformers.js) JavaScript library from [NPM](https://www.npmjs.com/package/@huggingface/transformers) using: ```bash npm i @huggingface/transformers ``` **Example:** Perform mask generation with `Xenova/sam-vit-base`. ```js import { SamModel, AutoProcessor, RawImage } from "@huggingface/transformers"; // Load model and processor const model = await SamModel.from_pretrained("Xenova/sam-vit-base"); const processor = await AutoProcessor.from_pretrained("Xenova/sam-vit-base"); // Prepare image and input points const img_url = "https://huggingface.co/datasets/Xenova/transformers.js-docs/resolve/main/corgi.jpg"; const raw_image = await RawImage.read(img_url); const input_points = [[[340, 250]]]; // Process inputs and perform mask generation const inputs = await processor(raw_image, { input_points }); const outputs = await model(inputs); // Post-process masks const masks = await processor.post_process_masks(outputs.pred_masks, inputs.original_sizes, inputs.reshaped_input_sizes); console.log(masks); // [ // Tensor { // dims: [ 1, 3, 410, 614 ], // type: 'bool', // data: Uint8Array(755220) [ ... ], // size: 755220 // } // ] const scores = outputs.iou_scores; console.log(scores); // Tensor { // dims: [ 1, 1, 3 ], // type: 'float32', // data: Float32Array(3) [ // 0.9466127157211304, // 0.9890615344047546, // 0.8316894769668579 // ], // size: 3 // } ``` You can then visualize the generated mask with: ```js const image = RawImage.fromTensor(masks[0][0].mul(255)); image.save('mask.png'); ``` ![image/png](https://cdn-uploads.huggingface.co/production/uploads/61b253b7ac5ecaae3d1efe0c/poSjt_XvMNO7ZluYFX28e.png) Next, select the channel with the highest IoU score, which in this case is the second (green) channel. Intersecting this with the original image gives us an isolated version of the subject: ![image/gif](https://cdn-uploads.huggingface.co/production/uploads/61b253b7ac5ecaae3d1efe0c/uQHu_vleyldfeEX2Q4d0G.gif) ## Demo We've also got an online demo, which you can try out [here](https://huggingface.co/spaces/Xenova/segment-anything-web). <video controls autoplay src="https://cdn-uploads.huggingface.co/production/uploads/61b253b7ac5ecaae3d1efe0c/Y0wAOw6hz9rWpwiuMoz2A.mp4"></video> --- Note: Having a separate repo for ONNX weights is intended to be a temporary solution until WebML gains more traction. If you would like to make your models web-ready, we recommend converting to ONNX using [🤗 Optimum](https://huggingface.co/docs/optimum/index) and structuring your repo like this one (with ONNX weights located in a subfolder named `onnx`).