Search is not available for this dataset
pipeline_tag
stringclasses 48
values | library_name
stringclasses 205
values | text
stringlengths 0
18.3M
| metadata
stringlengths 2
1.07B
| id
stringlengths 5
122
| last_modified
null | tags
sequencelengths 1
1.84k
| sha
null | created_at
stringlengths 25
25
|
---|---|---|---|---|---|---|---|---|
null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
| {"library_name": "transformers", "tags": []} | twodigit/Meta-Llama-3-8B-Instruct-koconv2_4327k-sft-lora-120000 | null | [
"transformers",
"safetensors",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T10:24:02+00:00 |
null | null | {"license": "openrail"} | de0nis2011/KaiAngel | null | [
"license:openrail",
"region:us"
] | null | 2024-05-01T10:26:37+00:00 |
|
null | transformers |
# Uploaded model
- **Developed by:** srbdtwentyfour
- **License:** apache-2.0
- **Finetuned from model :** unsloth/llama-3-8b-Instruct-bnb-4bit
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
| {"language": ["en"], "license": "apache-2.0", "tags": ["text-generation-inference", "transformers", "unsloth", "llama", "trl"], "base_model": "unsloth/llama-3-8b-Instruct-bnb-4bit"} | srbdtwentyfour/mystery-llama-3-8b | null | [
"transformers",
"safetensors",
"text-generation-inference",
"unsloth",
"llama",
"trl",
"en",
"base_model:unsloth/llama-3-8b-Instruct-bnb-4bit",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T10:26:41+00:00 |
text-generation | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {"library_name": "transformers", "tags": []} | gen-bi/llama-2-ko-juno-7b | null | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"4-bit",
"region:us"
] | null | 2024-05-01T10:28:12+00:00 |
text-generation | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {"library_name": "transformers", "tags": []} | rainerberger/planetn6 | null | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T10:29:05+00:00 |
reinforcement-learning | ml-agents |
# **ppo** Agent playing **Huggy**
This is a trained model of a **ppo** agent playing **Huggy**
using the [Unity ML-Agents Library](https://github.com/Unity-Technologies/ml-agents).
## Usage (with ML-Agents)
The Documentation: https://unity-technologies.github.io/ml-agents/ML-Agents-Toolkit-Documentation/
We wrote a complete tutorial to learn to train your first agent using ML-Agents and publish it to the Hub:
- A *short tutorial* where you teach Huggy the Dog 🐶 to fetch the stick and then play with him directly in your
browser: https://huggingface.co/learn/deep-rl-course/unitbonus1/introduction
- A *longer tutorial* to understand how works ML-Agents:
https://huggingface.co/learn/deep-rl-course/unit5/introduction
### Resume the training
```bash
mlagents-learn <your_configuration_file_path.yaml> --run-id=<run_id> --resume
```
### Watch your Agent play
You can watch your agent **playing directly in your browser**
1. If the environment is part of ML-Agents official environments, go to https://huggingface.co/unity
2. Step 1: Find your model_id: Chhabi/ppo-Huggy
3. Step 2: Select your *.nn /*.onnx file
4. Click on Watch the agent play 👀
| {"library_name": "ml-agents", "tags": ["Huggy", "deep-reinforcement-learning", "reinforcement-learning", "ML-Agents-Huggy"]} | Chhabi/ppo-Huggy | null | [
"ml-agents",
"tensorboard",
"onnx",
"Huggy",
"deep-reinforcement-learning",
"reinforcement-learning",
"ML-Agents-Huggy",
"region:us"
] | null | 2024-05-01T10:29:19+00:00 |
reinforcement-learning | null |
# **Q-Learning** Agent playing1 **FrozenLake-v1**
This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** .
## Usage
```python
model = load_from_hub(repo_id="ArnavModanwal/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl")
# Don't forget to check if you need to add additional attributes (is_slippery=False etc)
env = gym.make(model["env_id"])
```
| {"tags": ["FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "q-FrozenLake-v1-4x4-noSlippery", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "FrozenLake-v1-4x4-no_slippery", "type": "FrozenLake-v1-4x4-no_slippery"}, "metrics": [{"type": "mean_reward", "value": "1.00 +/- 0.00", "name": "mean_reward", "verified": false}]}]}]} | ArnavModanwal/q-FrozenLake-v1-4x4-noSlippery | null | [
"FrozenLake-v1-4x4-no_slippery",
"q-learning",
"reinforcement-learning",
"custom-implementation",
"model-index",
"region:us"
] | null | 2024-05-01T10:29:22+00:00 |
null | null | {} | princo/Face_recognition_exam | null | [
"region:us"
] | null | 2024-05-01T10:29:40+00:00 |
|
null | transformers | {} | rancelyndar/segformer-b5-asbestos-full-imagesV1-g | null | [
"transformers",
"tensorboard",
"safetensors",
"segformer",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T10:30:01+00:00 |
|
null | null | {"license": "mit"} | Achillektf/F_News | null | [
"license:mit",
"region:us"
] | null | 2024-05-01T10:30:08+00:00 |
|
null | null | {} | davidrockefeller/papipum-kaggle-SDXL | null | [
"region:us"
] | null | 2024-05-01T10:30:35+00:00 |
|
text-generation | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {"library_name": "transformers", "tags": []} | RefalMachine/ruadapt_llama3_part1-2_vo_3e4_bs256 | null | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T10:31:34+00:00 |
null | null | {} | miittnnss/portrait-ddim | null | [
"region:us"
] | null | 2024-05-01T10:32:30+00:00 |
|
null | null | {} | Selma323/whisper-small-dv | null | [
"region:us"
] | null | 2024-05-01T10:32:33+00:00 |
|
null | null | {} | Kaizu07/ddp_llama_bn_finetuned_v0.2_q4_0.gguf | null | [
"gguf",
"region:us"
] | null | 2024-05-01T10:32:37+00:00 |
|
null | null | {} | reevan/gemma_kan_rom_4bit | null | [
"region:us"
] | null | 2024-05-01T10:32:56+00:00 |
|
null | null | {} | davidrockefeller/papipum-easyphoto-512 | null | [
"region:us"
] | null | 2024-05-01T10:33:59+00:00 |
|
null | transformers |
# Uploaded model
- **Developed by:** davanstrien
- **License:** apache-2.0
- **Finetuned from model :** unsloth/llama-3-8b-bnb-4bit
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
| {"language": ["en"], "license": "apache-2.0", "tags": ["text-generation-inference", "transformers", "unsloth", "llama", "gguf"], "base_model": "unsloth/llama-3-8b-bnb-4bit"} | davanstrien/LLama-3-dataset-tldr-gguf | null | [
"transformers",
"gguf",
"llama",
"text-generation-inference",
"unsloth",
"en",
"base_model:unsloth/llama-3-8b-bnb-4bit",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T10:34:36+00:00 |
text-generation | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {"library_name": "transformers", "tags": ["unsloth"]} | reevan/gemma_kan_rom_16bit | null | [
"transformers",
"pytorch",
"gemma",
"text-generation",
"unsloth",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T10:34:38+00:00 |
feature-extraction | transformers | {"license": "mit"} | akkipro/e5-base-v2-ov-int8 | null | [
"transformers",
"openvino",
"bert",
"feature-extraction",
"license:mit",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T10:36:02+00:00 |
|
reinforcement-learning | null |
# **Q-Learning** Agent playing1 **Taxi-v3**
This is a trained model of a **Q-Learning** agent playing **Taxi-v3** .
## Usage
```python
model = load_from_hub(repo_id="ArnavModanwal/Taxi-v3", filename="q-learning.pkl")
# Don't forget to check if you need to add additional attributes (is_slippery=False etc)
env = gym.make(model["env_id"])
```
| {"tags": ["Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "Taxi-v3", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "Taxi-v3", "type": "Taxi-v3"}, "metrics": [{"type": "mean_reward", "value": "7.56 +/- 2.71", "name": "mean_reward", "verified": false}]}]}]} | ArnavModanwal/Taxi-v3 | null | [
"Taxi-v3",
"q-learning",
"reinforcement-learning",
"custom-implementation",
"model-index",
"region:us"
] | null | 2024-05-01T10:36:05+00:00 |
text-classification | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# emotion-turkish19
This model is a fine-tuned version of [dbmdz/bert-base-turkish-cased](https://huggingface.co/dbmdz/bert-base-turkish-cased) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2663
- Accuracy: 0.9333
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| No log | 1.0 | 71 | 0.2038 | 0.9524 |
| No log | 2.0 | 142 | 0.2325 | 0.9333 |
| No log | 3.0 | 213 | 0.2663 | 0.9333 |
### Framework versions
- Transformers 4.40.1
- Pytorch 2.2.1+cu121
- Datasets 2.19.0
- Tokenizers 0.19.1
| {"license": "mit", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "dbmdz/bert-base-turkish-cased", "model-index": [{"name": "emotion-turkish19", "results": []}]} | asude55/emotion-turkish19 | null | [
"transformers",
"safetensors",
"bert",
"text-classification",
"generated_from_trainer",
"base_model:dbmdz/bert-base-turkish-cased",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T10:36:20+00:00 |
text-generation | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {"library_name": "transformers", "tags": []} | steve1989/fingpt-SA-bnb-4bits-finedtuned-financialphrasebank | null | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T10:36:29+00:00 |
null | mlx |
# GreenBitAI/01-Yi-9B-layer-mix-bpw-3.0-mlx
This quantized low-bit model was converted to MLX format from [`GreenBitAI/01-Yi-9B-layer-mix-bpw-3.0`]().
Refer to the [original model card](https://huggingface.co/GreenBitAI/01-Yi-9B-layer-mix-bpw-3.0) for more details on the model.
## Use with mlx
```bash
pip install gbx-lm
```
```python
from gbx_lm import load, generate
model, tokenizer = load("GreenBitAI/01-Yi-9B-layer-mix-bpw-3.0-mlx")
response = generate(model, tokenizer, prompt="hello", verbose=True)
```
| {"license": "apache-2.0", "tags": ["mlx"]} | GreenBitAI/01-Yi-9B-layer-mix-bpw-3.0-mlx | null | [
"mlx",
"safetensors",
"llama",
"license:apache-2.0",
"region:us"
] | null | 2024-05-01T10:37:04+00:00 |
text-generation | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {"library_name": "transformers", "tags": []} | sanchit42/Mistral-7b-4bit-finetune | null | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T10:37:25+00:00 |
text-generation | transformers | {} | Nattipon/openthaigpt-1.0.0-beta-7b-ckpt-hf-cancer5k-4ep | null | [
"transformers",
"pytorch",
"llama",
"text-generation",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T10:38:20+00:00 |
|
text-classification | setfit |
# SetFit with sentence-transformers/paraphrase-MiniLM-L6-v2
This is a [SetFit](https://github.com/huggingface/setfit) model that can be used for Text Classification. This SetFit model uses [sentence-transformers/paraphrase-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/paraphrase-MiniLM-L6-v2) as the Sentence Transformer embedding model. A [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance is used for classification.
The model has been trained using an efficient few-shot learning technique that involves:
1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning.
2. Training a classification head with features from the fine-tuned Sentence Transformer.
## Model Details
### Model Description
- **Model Type:** SetFit
- **Sentence Transformer body:** [sentence-transformers/paraphrase-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/paraphrase-MiniLM-L6-v2)
- **Classification head:** a [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance
- **Maximum Sequence Length:** 128 tokens
- **Number of Classes:** 75 classes
<!-- - **Training Dataset:** [Unknown](https://huggingface.co/datasets/unknown) -->
<!-- - **Language:** Unknown -->
<!-- - **License:** Unknown -->
### Model Sources
- **Repository:** [SetFit on GitHub](https://github.com/huggingface/setfit)
- **Paper:** [Efficient Few-Shot Learning Without Prompts](https://arxiv.org/abs/2209.11055)
- **Blogpost:** [SetFit: Efficient Few-Shot Learning Without Prompts](https://huggingface.co/blog/setfit)
### Model Labels
| Label | Examples |
|:------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| 9 | <ul><li>'What type of fabric is recommended for creating comfortable clothing that is resistant to wear and tear?'</li><li>'What type of fabric is best for creating garments with slight nubs and variations for a natural look?'</li><li>'Where can I buy durable cotton fabric in deep olive green for everyday wear?'</li></ul> |
| 43 | <ul><li>'What is a tightly woven fabric suitable for lightweight jackets and formal trousers?'</li><li>'What fabric is not ideal for garments requiring significant stretch or drape, such as knitwear or flowy dresses?'</li><li>'Which textile is best for garments that need a subtle texture and medium weight?'</li></ul> |
| 66 | <ul><li>'Searching for a dark gray textile with a soft texture and fine weave pattern suitable for making skirts and dresses.'</li><li>'What fabric type is recommended for making garments that need to maintain their shape while being comfortable and adaptable for different styles?'</li><li>'Which fabric is suitable for making clothes that maintain their shape but also provide comfort and flexibility?'</li></ul> |
| 22 | <ul><li>'What fabric has a raised texture and tight weave for garments that require strength and longevity?'</li><li>'What fabric is recommended for garments that require both comfort and resilience?'</li><li>'What is the best fabric for creating outerwear with a medium weight and good body?'</li></ul> |
| 5 | <ul><li>'What kind of textile is suitable for crafting lightweight summer dresses with a fluid drape and hint of elasticity?'</li><li>'What type of textile and weave is consistent with an interlocking loop structure and stretchable properties?'</li><li>'What fabric can I use to make soft loungewear that has a luxurious feel and good performance in apparel?'</li></ul> |
| 52 | <ul><li>'What fabric has moisture-wicking properties for sporty summer wear?'</li><li>'Where to find textiles suitable for people with sensitive skin for comfortable wear?'</li><li>'What are the best fabrics for moisture-wicking properties in sporty or casual summer wear?'</li></ul> |
| 67 | <ul><li>'What fabric is recommended for making durable clothing with a smooth, consistent grain?'</li><li>'Which fabric has a solid color resembling taupe and a moderate saturation?'</li><li>'What kind of textile is good for creating garments with a soft drape and gentle folds?'</li></ul> |
| 32 | <ul><li>'What fabric is best suited for creating clothing with a fine gauge knit and a smooth flow for ease of movement?'</li><li>'What fabric is ideal for making form-fitting leggings and sports tops with good stretch and flexibility?'</li><li>'What type of fabric is recommended for crafting garments with a consistent dark gray hue and a slight sheen on the surface?'</li></ul> |
| 53 | <ul><li>'Where can I find a high-quality textile ideal for making athletic wear with stretchability?'</li><li>'What textile is perfect for making garments that require both structure and elasticity?'</li><li>'Which fabric is ideal for creating athletic wear with strong saturation and even color distribution?'</li></ul> |
| 16 | <ul><li>'What fabric has a textured surface with visible loops and a cozy hand feel?'</li><li>'What fabric is best for making durable garments that have a mottled black, white, and gray appearance?'</li><li>'What type of fabric displays a mottled grayscale coloration with a melange effect?'</li></ul> |
| 4 | <ul><li>'Which fabric has a fine knit weave, smooth texture, and a slight sheen?'</li><li>'What is the most suitable fabric for creating clothing items for individuals with sensitive skin?'</li><li>'What fabric can I use for creating lightweight and breathable summer tops with a soft texture?'</li></ul> |
| 65 | <ul><li>'What type of fabric is this deep blue twill textile with a slight rough texture and medium-weight suitable for?'</li><li>'What fabric would be suitable for making comfortable and form-fitting jeans?'</li><li>'What type of fabric is ideal for making durable and form-fitting jeans?'</li></ul> |
| 55 | <ul><li>'What is the composition of the knit fabric with a fluid drape and some stretch?'</li><li>'What fabric would be best for making form-fitting dresses that require some stretch and elasticity?'</li><li>'What fabric is suitable for form-fitting clothing like t-shirts, leggings, and dresses?'</li></ul> |
| 12 | <ul><li>'Which textile is suitable for garments that need a delicate fall and a matte finish?'</li><li>'What fabric is recommended for creating linings in apparel due to its lightness and versatility?'</li><li>'What is a versatile fabric option for making shirts that are both comfortable and durable?'</li></ul> |
| 71 | <ul><li>'Which textile exhibits a striped pattern achieved through yarn dyeing for a sharp contrast?'</li><li>'What type of cotton fabric has a smooth texture and is suitable for making summer dresses?'</li><li>'Which fabric is suitable for making casual shirting with a soft hand feel and fluid drape?'</li></ul> |
| 25 | <ul><li>'What material is floppy with some flexibility but not significant stretch?'</li><li>'Which fabric is better for utility wear rather than structured silhouettes?'</li><li>'What textile has small colorful fibers and lacks a traditional woven or knitted structure?'</li></ul> |
| 6 | <ul><li>'What fabric is suitable for casual wear and layering in various climates with a subtle sheen and clean surface?'</li><li>'What fabric can I use to make moisture-wicking clothing suitable for people with sensitive skin and a versatile look?'</li><li>'What fabric can I use to create garments that have a neat finish and attention to detail in the textile processing?'</li></ul> |
| 20 | <ul><li>'What fabric is versatile for multi-seasonal use, durable, and maintains its shape over time?'</li><li>'What fabric is recommended for making leggings and casual wear with a balanced drape and consistent coloring?'</li><li>'Where can I find a fabric suitable for multi-seasonal use with a consistent hue and soft hand texture?'</li></ul> |
| 10 | <ul><li>'What type of cotton fabric is ideal for making casual shirts and trousers?'</li><li>'Which fabric has a soft drape and medium weight for making versatile garments?'</li><li>'What type of fabric is ideal for making versatile garments with good movement and flow?'</li></ul> |
| 0 | <ul><li>'Which fabric has a clean appearance with a subtle sheen from bamboo fibers?'</li><li>'Which fabric is ideal for making garments that need to maintain their shape but have some stretch?'</li><li>'What fabric is recommended for making garments with a clean and even black color without significant variations or patterns?'</li></ul> |
| 42 | <ul><li>'What fabric is suitable for making versatile dresses with a fluid drape and stretchy feel?'</li><li>'What type of knit fabric is recommended for creating garments that require a fluid drape and some degree of elasticity?'</li><li>'Where can I find a vibrant red fabric with high saturation for making eye-catching garments?'</li></ul> |
| 57 | <ul><li>'What type of fabric is light grey with a cool undertone and has a soft, fluid drape?'</li><li>'What material is best for making comfortable and durable clothing suitable for regular wear?'</li><li>'Which fabric offers a combination of comfort, durability, and stretch for versatile garment applications?'</li></ul> |
| 36 | <ul><li>'What fabric can I use to make comfortable and flexible activewear?'</li><li>'What type of fabric is best for making lightweight sweaters with a smooth texture?'</li><li>'What type of textile is best for making layering pieces for cooler climates?'</li></ul> |
| 37 | <ul><li>'What textile is smooth with fine threads and a gentle drape?'</li><li>'What is the best fabric for creating breathable and comfortable dresses for warm weather?'</li><li>'What type of fabric is best for creating lightweight blouses with a soft drape?'</li></ul> |
| 58 | <ul><li>"Which textile is lightweight and breathable, suitable for children's wear with a green and blue floral design?"</li><li>'Ideal textile for t-shirts that require a degree of stretchability'</li><li>'Which fabric is recommended for creating garments with moisture-wicking properties and a vibrant color palette?'</li></ul> |
| 56 | <ul><li>'What type of fabric is this medium grey textile with a smooth drape and slight stretch?'</li><li>'What is the best fabric for making light sweaters that are durable and long-lasting?'</li><li>'What type of fabric is ideal for making everyday wear garments with a smooth texture and solid color?'</li></ul> |
| 17 | <ul><li>'What fabric is textured with fine loops and suitable for creating garments that require some structural qualities?'</li><li>'What fabric exhibits a brushed or fleeced finish and would be perfect for crafting cozy winter clothing?'</li><li>'What fabric is recommended for fall and winter activewear due to its warmth and comfort?'</li></ul> |
| 72 | <ul><li>'What is a versatile cotton fabric with fine to medium thread count, perfect for creating breathable garments for warm climates?'</li><li>'What fabric is ideal for making blouses and dresses with a simple, unadorned aesthetic?'</li><li>'What fabric is suitable for creating durable and versatile garments without unique finishes?'</li></ul> |
| 54 | <ul><li>'Looking for a fabric suitable for making lightweight jackets with a soft drape.'</li><li>'What type of fabric is commonly used in t-shirts for a comfortable and breathable feel?'</li><li>'What kind of textile weave is ideal for crafting casual t-shirts with some stretchability?'</li></ul> |
| 59 | <ul><li>'Where can I find a knit fabric with a slightly textured surface and fine, soft feel that is comfortable for casual wear?'</li><li>'What fabric is versatile and comfortable for casual wear?'</li><li>'What knit fabric is ideal for making dresses that require a bit of stretch and versatility in styling?'</li></ul> |
| 60 | <ul><li>'What fabric would be suitable for making t-shirts that conform well to body shapes and have vibrant hues?'</li><li>'Where can I find a jersey knit fabric with a smooth texture and fine knit structure suitable for t-shirts?'</li><li>'What type of fabric is this deep purple floral patterned material made of?'</li></ul> |
| 1 | <ul><li>'What is the best fabric for making clothing with moisture-wicking properties?'</li><li>'What type of fabric would be recommended for creating structured garments that also offer stretch and flexibility?'</li><li>'What is the best fabric for making clothing with moisture-wicking properties?'</li></ul> |
| 47 | <ul><li>'What type of textile is ideal for making spring and summer leggings with a smooth texture and stretchability?'</li><li>'Which fabric is lightweight and ideal for creating leggings that maintain their shape and offer flexibility?'</li><li>'What fabric composition is suitable for creating lightweight jackets that allow for movement and breathability?'</li></ul> |
| 28 | <ul><li>'What fabric is suitable for making blouses, dresses, skirts, and lightweight jackets?'</li><li>'What fabric with a smooth surface and medium weight is suitable for structured garments?'</li><li>'What fabric is durable and likely to maintain its color and shape well?'</li></ul> |
| 13 | <ul><li>'Which fabric is recommended for casual loungewear that needs to be both comfortable and resilient?'</li><li>'What is the best fabric blend for making soft and durable lightweight sweaters?'</li><li>'What type of fabric offers a good balance between performance and aesthetics for everyday wear?'</li></ul> |
| 26 | <ul><li>'What fabric has a plain weave pattern, smooth surface, and fine thread count with a slight sheen?'</li><li>'Is there a fabric with moderate strength and a smooth finish ideal for creating garments with soft silhouettes?'</li><li>'What fabric is 100% Rayon, lightweight, and ideal for creating garments with soft silhouettes?'</li></ul> |
| 15 | <ul><li>'What knit fabric would be suitable for making cozy apparel with warmth without excessive bulk?'</li><li>'Which fabric is best for creating casual wear with an understated aesthetic and versatile appeal?'</li><li>'What type of fabric is characterized by a melange of earthy tones with a heathered effect?'</li></ul> |
| 50 | <ul><li>'Where can I find a vibrant blue fabric with consistent dye saturation for t-shirts and activewear?'</li><li>'What fabric is best for creating clothing with a consistent, even dye and some stretchability for comfort and durability?'</li><li>'Where can I find a knit fabric with vibrant blue color and a smooth, fine texture?'</li></ul> |
| 24 | <ul><li>'What type of polyester fabric offers a comfortable fit with a moderate drape for daily wear?'</li><li>'What fabric has a textured surface and slight elasticity for comfortable fit?'</li><li>'What type of textile is recommended for garments that require consistent saturation and evenness in color?'</li></ul> |
| 29 | <ul><li>'Which fabric is ideal for creating garments that can withstand regular wear and maintain their texture over time?'</li><li>'What type of fabric has a consistent grey hue with a subtle mottled appearance?'</li><li>'What polyester textile has a micro crinkle texture and fine threads?'</li></ul> |
| 44 | <ul><li>'What knit textile is suitable for creating casual dresses with a fluid drape and soft texture?'</li><li>"I'm searching for a jersey knit fabric with durable, wrinkle-resistant properties for everyday wear, do you have any options?"</li><li>'What type of knit fabric is recommended for everyday apparel due to its comfort and ease of movement?'</li></ul> |
| 38 | <ul><li>'What fabric is best for creating blouses with a clean and crisp appearance?'</li><li>'What type of fabric provides a combination of durability and practicality for everyday wear garments?'</li><li>"I'm looking for a fabric with a clean and crisp appearance that is durable and easy to care for, any suggestions?"</li></ul> |
| 23 | <ul><li>'What fabric is appropriate for garments that require a hint of texture in the surface?'</li><li>'What type of fabric is suitable for creating structured jackets and trousers with a professional look?'</li><li>'What fabric is suitable for making medium-weight garments with a hint of roughness in texture?'</li></ul> |
| 45 | <ul><li>'Interested in a fabric with stretch and recovery for making garments that require some elasticity and resilience?'</li><li>'Which fabric is recommended for creating durable clothing suitable for people with sensitive skin, featuring a smooth texture and vibrant blue color with white dots?'</li><li>'What fabric is recommended for making polka dot clothing with a smooth surface and vibrant color?'</li></ul> |
| 31 | <ul><li>'What type of knit textile is recommended for creating layering pieces in solid, dark colors?'</li><li>'What is a versatile fabric for creating garments with a matte finish and uniform color?'</li><li>'Which fabric is suitable for activewear, leggings, and fitted tops due to its stretchability?'</li></ul> |
| 19 | <ul><li>"What type of fabric is ideal for making playful children's wear with a vibrant speckled pattern?"</li><li>'Which fabric is suitable for crafting garments that can hide wear and minor soiling due to its unique speckled pattern?'</li><li>'What fabric offers good recovery and fit due to elastane content?'</li></ul> |
| 11 | <ul><li>'What is a medium weight textile with a soft drape for creating versatile garments?'</li><li>'What fabric is lightweight and breathable, perfect for making soft summer blouses?'</li><li>'Which fabric is suitable for making soft and comfortable shirts and blouses with a consistent light blue hue?'</li></ul> |
| 73 | <ul><li>'What type of fabric is suitable for apparel that requires both form and function?'</li><li>'Best fabric for creating statement pieces with a pop of color using a twill weave texture?'</li><li>'Which fabric has a slightly textured surface with medium fineness threads, ideal for structured garments?'</li></ul> |
| 64 | <ul><li>'What fabric would be best for making pants that maintain their shape while offering flexibility?'</li><li>'What fabric blend offers both comfort and durability for creating long-lasting clothing?'</li><li>'Which fabric is known for its simple yet durable qualities with no unique finishes?'</li></ul> |
| 35 | <ul><li>'What type of fabric is recommended for creating breathable and comfortable clothing for warm weather?'</li><li>'What fabric would be suitable for making lightweight sweaters with a ribbed texture and soft hand?'</li><li>'What type of fabric is best for making form-fitting t-shirts with a fluid drape?'</li></ul> |
| 21 | <ul><li>'What fabric blend offers durability and slight stretchability for structured yet comfortable dresses?'</li><li>'What fabric is durable yet versatile for various garment constructions?'</li><li>'What type of cloth is versatile for various seasons due to its weight and composition?'</li></ul> |
| 74 | <ul><li>'Need medium weight cotton fabric for creating casual shirts with a balanced color scheme?'</li><li>'Looking for plain weave cotton fabric with a fine thread count and even color distribution?'</li><li>'Which textile is versatile for various seasons like spring and summer due to its lightness?'</li></ul> |
| 3 | <ul><li>'Looking for a fabric for casual apparel applications in mild to warm climates with consistent dyeing?'</li><li>'Which fabric blend is recommended for creating apparel with both breathability and a gentle flow?'</li><li>'Where can I purchase a bamboo-spandex blend fabric suitable for all-season clothing with moisture-wicking properties?'</li></ul> |
| 8 | <ul><li>'What type of fabric is ideal for creating form-fitting tops with a fluid drape?'</li><li>'What fabric composition combines bamboo and Pret fibers for eco-friendly benefits?'</li><li>'What fabric can I use to make elegant and comfortable cardigans with stretch properties?'</li></ul> |
| 18 | <ul><li>'What fabric is recommended for making lightweight garments with a smooth flow and gentle folds?'</li><li>'What type of knit fabric is ideal for creating dresses with moderate stretchability?'</li><li>'What textile composition includes elastane and bamboo for stretchability and comfort in casual apparel?'</li></ul> |
| 49 | <ul><li>'What is the ideal textile for crafting activewear with moderate weight and stretch?'</li><li>'Where can I find a jersey knit textile with a soft texture and fine fibers for casual wear?'</li><li>'What is the recommended material for making activewear that allows for ease of movement?'</li></ul> |
| 27 | <ul><li>'What is the recommended fabric for creating spring and summer wear with a focus on breathability?'</li><li>'Which textile is recommended for creating blouses, skirts, and other apparel due to its natural sheen and uniform texture?'</li><li>'What type of fabric has a consistent coloration and high level of saturation for apparel applications?'</li></ul> |
| 63 | <ul><li>'Which fabric has a plain weave construction and a fine thread count for a smooth texture?'</li><li>'What fabric is durable and versatile for everyday wear?'</li><li>'What fabric can be used to make form-fitting clothing like dresses, thanks to its stretchability?'</li></ul> |
| 61 | <ul><li>'What fabric can I use to make casual dresses with a smooth texture and a lightweight feel?'</li><li>'What is a fabric with a tight structure and smooth drape ideal for making casual summer outfits?'</li><li>'What type of fabric is lightweight, breathable, and suitable for layering in variable climates?'</li></ul> |
| 34 | <ul><li>'What fabric is a periwinkle blue color with medium saturation and no visible defects?'</li><li>'What fabric has a soft and smooth texture with fine threads and a knit pattern?'</li><li>'Searching for a fabric that is durable, breathable, and suitable for people with sensitive skin, any options?'</li></ul> |
| 30 | <ul><li>'Are there any fabrics with a simple weave pattern that offer stretchability for semi-fitted garments?'</li><li>'What is the best fabric for creating garments with a good balance of structure and elasticity?'</li><li>'What fabric is suitable for creating garments that require good stretchability and resilience?'</li></ul> |
| 7 | <ul><li>'What type of fabric is commonly used in casual wear, loungewear, and active wear due to its durability and performance?'</li><li>'What type of fabric is suitable for creating comfortable loungewear and lightweight sweaters with a fine, smooth texture and good fabric care?'</li><li>'What is the best fabric for making active wear that offers breathability and performance?'</li></ul> |
| 14 | <ul><li>'What material provides a fluid drape and enough structure for t-shirts and lounge pants?'</li><li>'What fabric should I choose for producing clothing with good colorfastness and ease of care in a polyester composition?'</li><li>'What is the best material for creating casual dresses with a medium weight drape and a mix of darker and lighter grey tones?'</li></ul> |
| 48 | <ul><li>'What is the best fabric for making comfortable and stretchy t-shirts with a casual aesthetic?'</li><li>'Where can I buy a knit fabric that is versatile in styling and functional qualities for a range of clothing?'</li><li>'What type of fabric is durable and suitable for everyday wear with a casual aesthetic?'</li></ul> |
| 2 | <ul><li>'Which fabric contains bamboo and Spandex for creating comfortable casual dresses?'</li><li>'What fabric has a fluid drape and slight elasticity, suitable for summer dresses?'</li><li>'What is the recommended fabric for creating draped garments like dresses or tunics?'</li></ul> |
| 46 | <ul><li>'Which fabric is ideal for creating lightweight sweaters with a comfortable and breathable feel?'</li><li>'What type of fabric is ideal for making casual t-shirts with a vibrant striped pattern?'</li><li>'What is the recommended textile for making versatile garments that can be layered in cooler climates?'</li></ul> |
| 51 | <ul><li>'What knit fabric is versatile for use in various seasons and holds its shape well?'</li><li>'What type of fabric is recommended for creating casual tops with a gentle, soft drape?'</li><li>'What fabric is suitable for making lightweight and comfortable casual tops for everyday wear?'</li></ul> |
| 39 | <ul><li>'What fabric would be recommended for making moisture-wicking blouses suitable for warm climates?'</li><li>'What fabric would be apt for creating garments that require a fine, even weave structure?'</li><li>'What is a suitable fabric for creating drapery in light jackets with a slight sheen?'</li></ul> |
| 70 | <ul><li>'What type of cotton fabric is ideal for making shirts and blouses with a soft drape?'</li><li>'What textile has a slightly textured surface with a fine yet distinct weave?'</li><li>'Which cotton fabric is versatile and suitable for both menswear and womenswear?'</li></ul> |
| 68 | <ul><li>'Which fabric is breathable and soft to the touch, suitable for creating comfortable dresses?'</li><li>'Which fabric is recommended for making year-round garments with high color saturation?'</li><li>'What fabric can be used for making shirts, pants, and dresses that require a smooth drape and a hint of elasticity?'</li></ul> |
| 40 | <ul><li>'Which fabric is ideal for creating spring and summer collections with a soft touch and lightweight feel?'</li><li>'What textile is known for its easy care and durability in garment construction?'</li><li>'What type of fabric is best suited for creating blouses with a flowing drape and smooth texture?'</li></ul> |
| 69 | <ul><li>'Which fabric is durable, resilient, and has a slight give due to the Spandex content?'</li><li>'What fabric has a consistent charcoal gray hue with a matte finish and a twill weave pattern?'</li><li>'What fabric is recommended for making form-fitting jackets that are both durable and breathable?'</li></ul> |
| 33 | <ul><li>'What fabric would be suitable for creating draped skirts with a smooth surface and stretchability?'</li><li>'What is the best textile for creating draped skirts with a subtle iridescence?'</li><li>'Searching for a fabric with a smooth texture and slight shimmer effect for draped skirts?'</li></ul> |
| 41 | <ul><li>'What fabric has a soft drape and gentle folds, making it perfect for creating flowy and comfortable spring and summer dresses?'</li><li>'What type of knit fabric offers good resistance to wrinkles and shrinkage for practical everyday wear?'</li><li>'Searching for a polyester knit fabric with a consistent hue and saturation for making versatile and adaptable garments.'</li></ul> |
| 62 | <ul><li>'Which fabric is versatile and suitable for creating durable garments for everyday wear?'</li><li>'What fabric is suitable for making casual wear like t-shirts, dresses, and tops?'</li><li>'What fabric is known for its stable weave with a small percentage of elastane for comfort and durability?'</li></ul> |
## Evaluation
### Metrics
| Label | Accuracy |
|:--------|:---------|
| **all** | 0.3463 |
## Uses
### Direct Use for Inference
First install the SetFit library:
```bash
pip install setfit
```
Then you can load this model and run inference.
```python
from setfit import SetFitModel
# Download from the 🤗 Hub
model = SetFitModel.from_pretrained("Jazielinho/fabric_model")
# Run inference
preds = model("What fabric has a comfortable feel and is suitable for people with sensitive skin?")
```
<!--
### Downstream Use
*List how someone could finetune this model on their own dataset.*
-->
<!--
### Out-of-Scope Use
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
-->
<!--
## Bias, Risks and Limitations
*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
-->
<!--
### Recommendations
*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
-->
## Training Details
### Training Set Metrics
| Training set | Min | Median | Max |
|:-------------|:----|:--------|:----|
| Word count | 7 | 15.4858 | 30 |
| Label | Training Sample Count |
|:------|:----------------------|
| 0 | 39 |
| 1 | 40 |
| 2 | 41 |
| 3 | 32 |
| 4 | 37 |
| 5 | 33 |
| 6 | 36 |
| 7 | 40 |
| 8 | 30 |
| 9 | 36 |
| 10 | 42 |
| 11 | 38 |
| 12 | 39 |
| 13 | 43 |
| 14 | 41 |
| 15 | 41 |
| 16 | 35 |
| 17 | 42 |
| 18 | 40 |
| 19 | 43 |
| 20 | 44 |
| 21 | 36 |
| 22 | 37 |
| 23 | 40 |
| 24 | 44 |
| 25 | 42 |
| 26 | 41 |
| 27 | 38 |
| 28 | 41 |
| 29 | 46 |
| 30 | 41 |
| 31 | 38 |
| 32 | 40 |
| 33 | 39 |
| 34 | 41 |
| 35 | 44 |
| 36 | 45 |
| 37 | 40 |
| 38 | 37 |
| 39 | 44 |
| 40 | 39 |
| 41 | 42 |
| 42 | 36 |
| 43 | 43 |
| 44 | 42 |
| 45 | 37 |
| 46 | 41 |
| 47 | 44 |
| 48 | 36 |
| 49 | 40 |
| 50 | 43 |
| 51 | 44 |
| 52 | 39 |
| 53 | 38 |
| 54 | 38 |
| 55 | 43 |
| 56 | 41 |
| 57 | 44 |
| 58 | 40 |
| 59 | 41 |
| 60 | 35 |
| 61 | 43 |
| 62 | 41 |
| 63 | 43 |
| 64 | 37 |
| 65 | 41 |
| 66 | 36 |
| 67 | 38 |
| 68 | 42 |
| 69 | 41 |
| 70 | 39 |
| 71 | 43 |
| 72 | 34 |
| 73 | 40 |
| 74 | 41 |
### Training Hyperparameters
- batch_size: (256, 256)
- num_epochs: (20, 20)
- max_steps: -1
- sampling_strategy: undersampling
- body_learning_rate: (2e-05, 1e-05)
- head_learning_rate: 0.01
- loss: CosineSimilarityLoss
- distance_metric: cosine_distance
- margin: 0.25
- end_to_end: False
- use_amp: False
- warmup_proportion: 0.1
- seed: 42
- eval_max_steps: -1
- load_best_model_at_end: True
### Training Results
| Epoch | Step | Training Loss | Validation Loss |
|:-------:|:-------:|:-------------:|:---------------:|
| 0.0000 | 1 | 0.2732 | - |
| 0.0015 | 50 | 0.2545 | - |
| 0.0029 | 100 | 0.2538 | - |
| 0.0044 | 150 | 0.2633 | - |
| 0.0058 | 200 | 0.2598 | - |
| 0.0073 | 250 | 0.2624 | - |
| 0.0087 | 300 | 0.2537 | - |
| 0.0102 | 350 | 0.2592 | - |
| 0.0116 | 400 | 0.2475 | - |
| 0.0131 | 450 | 0.2483 | - |
| 0.0145 | 500 | 0.2418 | - |
| 0.0160 | 550 | 0.2403 | - |
| 0.0174 | 600 | 0.2386 | - |
| 0.0189 | 650 | 0.2542 | - |
| 0.0203 | 700 | 0.237 | - |
| 0.0218 | 750 | 0.2423 | - |
| 0.0232 | 800 | 0.2421 | - |
| 0.0247 | 850 | 0.2409 | - |
| 0.0261 | 900 | 0.2453 | - |
| 0.0276 | 950 | 0.2404 | - |
| 0.0290 | 1000 | 0.2418 | - |
| 0.0305 | 1050 | 0.2454 | - |
| 0.0319 | 1100 | 0.2446 | - |
| 0.0001 | 1 | 0.2471 | - |
| 0.0058 | 50 | 0.2375 | - |
| 0.0116 | 100 | 0.2351 | - |
| 0.0174 | 150 | 0.2406 | - |
| 0.0232 | 200 | 0.2382 | - |
| 0.0290 | 250 | 0.2374 | - |
| 0.0000 | 1 | 0.2515 | - |
| 0.0007 | 50 | 0.2335 | - |
| 0.0015 | 100 | 0.229 | - |
| 0.0022 | 150 | 0.2387 | - |
| 0.0029 | 200 | 0.2209 | - |
| 0.0036 | 250 | 0.2367 | - |
| 0.0044 | 300 | 0.2521 | - |
| 0.0051 | 350 | 0.239 | - |
| 0.0058 | 400 | 0.2405 | - |
| 0.0065 | 450 | 0.2541 | - |
| 0.0073 | 500 | 0.2308 | - |
| 0.0080 | 550 | 0.2381 | - |
| 0.0087 | 600 | 0.2456 | - |
| 0.0094 | 650 | 0.2301 | - |
| 0.0102 | 700 | 0.2486 | - |
| 0.0109 | 750 | 0.2243 | - |
| 0.0116 | 800 | 0.2399 | - |
| 0.0123 | 850 | 0.2341 | - |
| 0.0131 | 900 | 0.2417 | - |
| 0.0138 | 950 | 0.215 | - |
| 0.0145 | 1000 | 0.2264 | - |
| 0.0152 | 1050 | 0.2161 | - |
| 0.0160 | 1100 | 0.2273 | - |
| 0.0167 | 1150 | 0.2345 | - |
| 0.0174 | 1200 | 0.2302 | - |
| 0.0181 | 1250 | 0.2337 | - |
| 0.0189 | 1300 | 0.2278 | - |
| 0.0196 | 1350 | 0.2345 | - |
| 0.0203 | 1400 | 0.2323 | - |
| 0.0210 | 1450 | 0.2371 | - |
| 0.0218 | 1500 | 0.2217 | - |
| 0.0225 | 1550 | 0.2282 | - |
| 0.0232 | 1600 | 0.224 | - |
| 0.0239 | 1650 | 0.2346 | - |
| 0.0247 | 1700 | 0.2087 | - |
| 0.0254 | 1750 | 0.2299 | - |
| 0.0261 | 1800 | 0.2154 | - |
| 0.0268 | 1850 | 0.2108 | - |
| 0.0276 | 1900 | 0.216 | - |
| 0.0283 | 1950 | 0.2128 | - |
| 0.0290 | 2000 | 0.2083 | - |
| 0.0297 | 2050 | 0.2053 | - |
| 0.0305 | 2100 | 0.2265 | - |
| 0.0312 | 2150 | 0.2245 | - |
| 0.0319 | 2200 | 0.2036 | - |
| 0.0326 | 2250 | 0.2192 | - |
| 0.0334 | 2300 | 0.2259 | - |
| 0.0341 | 2350 | 0.2038 | - |
| 0.0348 | 2400 | 0.2129 | - |
| 0.0355 | 2450 | 0.2029 | - |
| 0.0363 | 2500 | 0.1883 | - |
| 0.0370 | 2550 | 0.187 | - |
| 0.0377 | 2600 | 0.2083 | - |
| 0.0384 | 2650 | 0.2138 | - |
| 0.0392 | 2700 | 0.2057 | - |
| 0.0399 | 2750 | 0.2134 | - |
| 0.0406 | 2800 | 0.2008 | - |
| 0.0413 | 2850 | 0.2018 | - |
| 0.0421 | 2900 | 0.2226 | - |
| 0.0428 | 2950 | 0.1815 | - |
| 0.0435 | 3000 | 0.1943 | - |
| 0.0442 | 3050 | 0.1926 | - |
| 0.0450 | 3100 | 0.1877 | - |
| 0.0457 | 3150 | 0.1764 | - |
| 0.0464 | 3200 | 0.2021 | - |
| 0.0471 | 3250 | 0.2071 | - |
| 0.0479 | 3300 | 0.1832 | - |
| 0.0486 | 3350 | 0.1714 | - |
| 0.0493 | 3400 | 0.1914 | - |
| 0.0500 | 3450 | 0.1749 | - |
| 0.0508 | 3500 | 0.1752 | - |
| 0.0515 | 3550 | 0.1829 | - |
| 0.0522 | 3600 | 0.175 | - |
| 0.0529 | 3650 | 0.1752 | - |
| 0.0537 | 3700 | 0.1973 | - |
| 0.0544 | 3750 | 0.1866 | - |
| 0.0551 | 3800 | 0.156 | - |
| 0.0558 | 3850 | 0.1923 | - |
| 0.0566 | 3900 | 0.1683 | - |
| 0.0573 | 3950 | 0.1642 | - |
| 0.0580 | 4000 | 0.1705 | - |
| 0.0587 | 4050 | 0.174 | - |
| 0.0595 | 4100 | 0.1609 | - |
| 0.0602 | 4150 | 0.17 | - |
| 0.0609 | 4200 | 0.1843 | - |
| 0.0616 | 4250 | 0.1855 | - |
| 0.0624 | 4300 | 0.1385 | - |
| 0.0631 | 4350 | 0.1765 | - |
| 0.0638 | 4400 | 0.1873 | - |
| 0.0645 | 4450 | 0.1654 | - |
| 0.0653 | 4500 | 0.1912 | - |
| 0.0660 | 4550 | 0.1533 | - |
| 0.0667 | 4600 | 0.1759 | - |
| 0.0674 | 4650 | 0.154 | - |
| 0.0682 | 4700 | 0.147 | - |
| 0.0689 | 4750 | 0.161 | - |
| 0.0696 | 4800 | 0.1603 | - |
| 0.0703 | 4850 | 0.1529 | - |
| 0.0711 | 4900 | 0.1538 | - |
| 0.0718 | 4950 | 0.1487 | - |
| 0.0725 | 5000 | 0.1593 | - |
| 0.0732 | 5050 | 0.1491 | - |
| 0.0740 | 5100 | 0.1389 | - |
| 0.0747 | 5150 | 0.1132 | - |
| 0.0754 | 5200 | 0.1622 | - |
| 0.0761 | 5250 | 0.1628 | - |
| 0.0769 | 5300 | 0.1598 | - |
| 0.0776 | 5350 | 0.1362 | - |
| 0.0783 | 5400 | 0.1637 | - |
| 0.0790 | 5450 | 0.1352 | - |
| 0.0798 | 5500 | 0.1523 | - |
| 0.0805 | 5550 | 0.1604 | - |
| 0.0812 | 5600 | 0.1534 | - |
| 0.0819 | 5650 | 0.1206 | - |
| 0.0827 | 5700 | 0.1331 | - |
| 0.0834 | 5750 | 0.1449 | - |
| 0.0841 | 5800 | 0.1376 | - |
| 0.0848 | 5850 | 0.1293 | - |
| 0.0856 | 5900 | 0.1258 | - |
| 0.0863 | 5950 | 0.1391 | - |
| 0.0870 | 6000 | 0.1678 | - |
| 0.0877 | 6050 | 0.1439 | - |
| 0.0885 | 6100 | 0.1329 | - |
| 0.0892 | 6150 | 0.1416 | - |
| 0.0899 | 6200 | 0.126 | - |
| 0.0906 | 6250 | 0.1072 | - |
| 0.0914 | 6300 | 0.1314 | - |
| 0.0921 | 6350 | 0.1282 | - |
| 0.0928 | 6400 | 0.1418 | - |
| 0.0935 | 6450 | 0.1418 | - |
| 0.0943 | 6500 | 0.1126 | - |
| 0.0950 | 6550 | 0.1118 | - |
| 0.0957 | 6600 | 0.1437 | - |
| 0.0964 | 6650 | 0.1265 | - |
| 0.0972 | 6700 | 0.1203 | - |
| 0.0979 | 6750 | 0.1267 | - |
| 0.0986 | 6800 | 0.11 | - |
| 0.0993 | 6850 | 0.1273 | - |
| 0.1001 | 6900 | 0.1253 | - |
| 0.1008 | 6950 | 0.1145 | - |
| 0.1015 | 7000 | 0.1054 | - |
| 0.1022 | 7050 | 0.1311 | - |
| 0.1030 | 7100 | 0.1238 | - |
| 0.1037 | 7150 | 0.0951 | - |
| 0.1044 | 7200 | 0.1187 | - |
| 0.1051 | 7250 | 0.1114 | - |
| 0.1059 | 7300 | 0.1038 | - |
| 0.1066 | 7350 | 0.1048 | - |
| 0.1073 | 7400 | 0.0965 | - |
| 0.1080 | 7450 | 0.1006 | - |
| 0.1088 | 7500 | 0.1273 | - |
| 0.1095 | 7550 | 0.12 | - |
| 0.1102 | 7600 | 0.1055 | - |
| 0.0001 | 1 | 0.1192 | - |
| 0.0029 | 50 | 0.1128 | - |
| 0.0057 | 100 | 0.0981 | - |
| 0.0021 | 1 | 0.1188 | - |
| 0.1040 | 50 | 0.1121 | - |
| 0.0021 | 1 | 0.1172 | - |
| 0.1040 | 50 | 0.1109 | - |
| 0.2079 | 100 | 0.0965 | - |
| 0.3119 | 150 | 0.1013 | - |
| 0.4158 | 200 | 0.1157 | - |
| 0.5198 | 250 | 0.1093 | - |
| 0.6237 | 300 | 0.1029 | - |
| 0.7277 | 350 | 0.0904 | - |
| 0.8316 | 400 | 0.1084 | - |
| 0.9356 | 450 | 0.1127 | - |
| **1.0** | **481** | **-** | **0.1883** |
| 1.0395 | 500 | 0.0853 | - |
| 1.1435 | 550 | 0.0907 | - |
| 1.2474 | 600 | 0.0814 | - |
| 1.3514 | 650 | 0.0967 | - |
| 1.4553 | 700 | 0.118 | - |
| 1.5593 | 750 | 0.0841 | - |
| 1.6632 | 800 | 0.0992 | - |
| 1.7672 | 850 | 0.0965 | - |
| 1.8711 | 900 | 0.092 | - |
| 1.9751 | 950 | 0.109 | - |
| 2.0 | 962 | - | 0.193 |
| 2.0790 | 1000 | 0.0847 | - |
| 2.1830 | 1050 | 0.0864 | - |
| 2.2869 | 1100 | 0.0843 | - |
| 2.3909 | 1150 | 0.0792 | - |
| 2.4948 | 1200 | 0.0808 | - |
| 2.5988 | 1250 | 0.0913 | - |
| 2.7027 | 1300 | 0.0848 | - |
| 2.8067 | 1350 | 0.0889 | - |
| 2.9106 | 1400 | 0.0673 | - |
| 3.0 | 1443 | - | 0.1983 |
| 3.0146 | 1450 | 0.0671 | - |
| 3.1185 | 1500 | 0.0643 | - |
| 3.2225 | 1550 | 0.0649 | - |
| 3.3264 | 1600 | 0.0827 | - |
| 3.4304 | 1650 | 0.0752 | - |
| 3.5343 | 1700 | 0.0785 | - |
| 3.6383 | 1750 | 0.0629 | - |
| 3.7422 | 1800 | 0.0726 | - |
| 3.8462 | 1850 | 0.0672 | - |
| 3.9501 | 1900 | 0.0704 | - |
| 4.0 | 1924 | - | 0.2015 |
| 4.0541 | 1950 | 0.0812 | - |
| 4.1580 | 2000 | 0.0709 | - |
| 4.2620 | 2050 | 0.0866 | - |
| 4.3659 | 2100 | 0.0747 | - |
| 4.4699 | 2150 | 0.0554 | - |
| 4.5738 | 2200 | 0.0636 | - |
| 4.6778 | 2250 | 0.0655 | - |
| 4.7817 | 2300 | 0.0562 | - |
| 4.8857 | 2350 | 0.0531 | - |
| 4.9896 | 2400 | 0.0518 | - |
| 5.0 | 2405 | - | 0.2056 |
| 5.0936 | 2450 | 0.0808 | - |
| 5.1975 | 2500 | 0.0571 | - |
| 5.3015 | 2550 | 0.066 | - |
| 5.4054 | 2600 | 0.071 | - |
| 5.5094 | 2650 | 0.0507 | - |
| 5.6133 | 2700 | 0.0603 | - |
| 5.7173 | 2750 | 0.0548 | - |
| 5.8212 | 2800 | 0.0714 | - |
| 5.9252 | 2850 | 0.0532 | - |
| 6.0 | 2886 | - | 0.208 |
| 6.0291 | 2900 | 0.0581 | - |
| 6.1331 | 2950 | 0.0663 | - |
| 6.2370 | 3000 | 0.0717 | - |
| 6.3410 | 3050 | 0.0549 | - |
| 6.4449 | 3100 | 0.0611 | - |
| 6.5489 | 3150 | 0.0515 | - |
| 6.6528 | 3200 | 0.0546 | - |
| 6.7568 | 3250 | 0.0406 | - |
| 6.8607 | 3300 | 0.0582 | - |
| 6.9647 | 3350 | 0.0565 | - |
| 7.0 | 3367 | - | 0.2176 |
| 7.0686 | 3400 | 0.0737 | - |
| 7.1726 | 3450 | 0.0554 | - |
| 7.2765 | 3500 | 0.0462 | - |
| 7.3805 | 3550 | 0.051 | - |
| 7.4844 | 3600 | 0.0441 | - |
| 7.5884 | 3650 | 0.0503 | - |
| 7.6923 | 3700 | 0.0531 | - |
| 7.7963 | 3750 | 0.0464 | - |
| 7.9002 | 3800 | 0.0443 | - |
| 8.0 | 3848 | - | 0.2234 |
| 8.0042 | 3850 | 0.0376 | - |
| 8.1081 | 3900 | 0.0542 | - |
| 8.2121 | 3950 | 0.0453 | - |
| 8.3160 | 4000 | 0.0448 | - |
| 8.4200 | 4050 | 0.0535 | - |
| 8.5239 | 4100 | 0.0645 | - |
| 8.6279 | 4150 | 0.0451 | - |
| 8.7318 | 4200 | 0.0472 | - |
| 8.8358 | 4250 | 0.0477 | - |
| 8.9397 | 4300 | 0.0327 | - |
| 9.0 | 4329 | - | 0.2272 |
| 9.0437 | 4350 | 0.0346 | - |
| 9.1476 | 4400 | 0.0435 | - |
| 9.2516 | 4450 | 0.0479 | - |
| 9.3555 | 4500 | 0.0508 | - |
| 9.4595 | 4550 | 0.0535 | - |
| 9.5634 | 4600 | 0.0631 | - |
| 9.6674 | 4650 | 0.0286 | - |
| 9.7713 | 4700 | 0.0564 | - |
| 9.8753 | 4750 | 0.0349 | - |
| 9.9792 | 4800 | 0.0487 | - |
| 10.0 | 4810 | - | 0.2288 |
| 10.0832 | 4850 | 0.0317 | - |
| 10.1871 | 4900 | 0.0546 | - |
| 10.2911 | 4950 | 0.0353 | - |
| 10.3950 | 5000 | 0.0437 | - |
| 10.4990 | 5050 | 0.056 | - |
| 10.6029 | 5100 | 0.0353 | - |
| 10.7069 | 5150 | 0.0304 | - |
| 10.8108 | 5200 | 0.0358 | - |
| 10.9148 | 5250 | 0.0481 | - |
| 11.0 | 5291 | - | 0.2282 |
| 11.0187 | 5300 | 0.0318 | - |
| 11.1227 | 5350 | 0.0373 | - |
| 11.2266 | 5400 | 0.0305 | - |
| 11.3306 | 5450 | 0.0443 | - |
| 11.4345 | 5500 | 0.0383 | - |
| 11.5385 | 5550 | 0.0425 | - |
| 11.6424 | 5600 | 0.039 | - |
| 11.7464 | 5650 | 0.0443 | - |
| 11.8503 | 5700 | 0.0503 | - |
| 11.9543 | 5750 | 0.0553 | - |
| 12.0 | 5772 | - | 0.2342 |
| 12.0582 | 5800 | 0.0362 | - |
| 12.1622 | 5850 | 0.0509 | - |
| 12.2661 | 5900 | 0.0337 | - |
| 12.3701 | 5950 | 0.0436 | - |
| 12.4740 | 6000 | 0.0462 | - |
| 12.5780 | 6050 | 0.034 | - |
| 12.6819 | 6100 | 0.0334 | - |
| 12.7859 | 6150 | 0.0365 | - |
| 12.8898 | 6200 | 0.047 | - |
| 12.9938 | 6250 | 0.0489 | - |
| 13.0 | 6253 | - | 0.2317 |
| 13.0977 | 6300 | 0.035 | - |
| 13.2017 | 6350 | 0.0412 | - |
| 13.3056 | 6400 | 0.0358 | - |
| 13.4096 | 6450 | 0.0366 | - |
| 13.5135 | 6500 | 0.0473 | - |
| 13.6175 | 6550 | 0.0481 | - |
| 13.7214 | 6600 | 0.0443 | - |
| 13.8254 | 6650 | 0.0454 | - |
| 13.9293 | 6700 | 0.0344 | - |
| 14.0 | 6734 | - | 0.2304 |
| 14.0333 | 6750 | 0.0327 | - |
| 14.1372 | 6800 | 0.0386 | - |
| 14.2412 | 6850 | 0.0503 | - |
| 14.3451 | 6900 | 0.0236 | - |
| 14.4491 | 6950 | 0.042 | - |
| 14.5530 | 7000 | 0.0405 | - |
| 14.6570 | 7050 | 0.0339 | - |
| 14.7609 | 7100 | 0.0435 | - |
| 14.8649 | 7150 | 0.0314 | - |
| 14.9688 | 7200 | 0.0263 | - |
| 15.0 | 7215 | - | 0.234 |
| 15.0728 | 7250 | 0.0369 | - |
| 15.1767 | 7300 | 0.0329 | - |
| 15.2807 | 7350 | 0.0366 | - |
| 15.3846 | 7400 | 0.0401 | - |
| 15.4886 | 7450 | 0.0321 | - |
| 15.5925 | 7500 | 0.0571 | - |
| 15.6965 | 7550 | 0.0353 | - |
| 15.8004 | 7600 | 0.0381 | - |
| 15.9044 | 7650 | 0.0347 | - |
| 16.0 | 7696 | - | 0.2334 |
| 16.0083 | 7700 | 0.0341 | - |
| 16.1123 | 7750 | 0.0276 | - |
| 16.2162 | 7800 | 0.0555 | - |
| 16.3202 | 7850 | 0.0338 | - |
| 16.4241 | 7900 | 0.0227 | - |
| 16.5281 | 7950 | 0.0256 | - |
| 16.6320 | 8000 | 0.0356 | - |
| 16.7360 | 8050 | 0.0413 | - |
| 16.8399 | 8100 | 0.032 | - |
| 16.9439 | 8150 | 0.0329 | - |
| 17.0 | 8177 | - | 0.2356 |
| 17.0478 | 8200 | 0.0382 | - |
| 17.1518 | 8250 | 0.0434 | - |
| 17.2557 | 8300 | 0.0411 | - |
| 17.3597 | 8350 | 0.0329 | - |
| 17.4636 | 8400 | 0.0388 | - |
| 17.5676 | 8450 | 0.0384 | - |
| 17.6715 | 8500 | 0.0306 | - |
| 17.7755 | 8550 | 0.0185 | - |
| 17.8794 | 8600 | 0.0357 | - |
| 17.9834 | 8650 | 0.0349 | - |
| 18.0 | 8658 | - | 0.2368 |
| 18.0873 | 8700 | 0.0515 | - |
| 18.1913 | 8750 | 0.0326 | - |
| 18.2952 | 8800 | 0.0367 | - |
| 18.3992 | 8850 | 0.0241 | - |
| 18.5031 | 8900 | 0.0313 | - |
| 18.6071 | 8950 | 0.0275 | - |
| 18.7110 | 9000 | 0.0378 | - |
| 18.8150 | 9050 | 0.0401 | - |
| 18.9189 | 9100 | 0.0285 | - |
| 19.0 | 9139 | - | 0.2347 |
| 19.0229 | 9150 | 0.0309 | - |
| 19.1268 | 9200 | 0.035 | - |
| 19.2308 | 9250 | 0.0415 | - |
| 19.3347 | 9300 | 0.0301 | - |
| 19.4387 | 9350 | 0.0293 | - |
| 19.5426 | 9400 | 0.0323 | - |
| 19.6466 | 9450 | 0.0342 | - |
| 19.7505 | 9500 | 0.0205 | - |
| 19.8545 | 9550 | 0.0331 | - |
| 19.9584 | 9600 | 0.0226 | - |
| 20.0 | 9620 | - | 0.237 |
* The bold row denotes the saved checkpoint.
### Framework Versions
- Python: 3.10.12
- SetFit: 1.0.3
- Sentence Transformers: 2.7.0
- Transformers: 4.40.1
- PyTorch: 2.2.1+cu121
- Datasets: 2.19.0
- Tokenizers: 0.19.1
## Citation
### BibTeX
```bibtex
@article{https://doi.org/10.48550/arxiv.2209.11055,
doi = {10.48550/ARXIV.2209.11055},
url = {https://arxiv.org/abs/2209.11055},
author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren},
keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
title = {Efficient Few-Shot Learning Without Prompts},
publisher = {arXiv},
year = {2022},
copyright = {Creative Commons Attribution 4.0 International}
}
```
<!--
## Glossary
*Clearly define terms in order to be accessible across audiences.*
-->
<!--
## Model Card Authors
*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
-->
<!--
## Model Card Contact
*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
--> | {"library_name": "setfit", "tags": ["setfit", "sentence-transformers", "text-classification", "generated_from_setfit_trainer"], "metrics": ["accuracy"], "base_model": "sentence-transformers/paraphrase-MiniLM-L6-v2", "widget": [{"text": "What fabric has a comfortable feel and is suitable for people with sensitive skin?"}, {"text": "What is the most recommended fabric for making outerwear that requires a blend of comfort and resilience?"}, {"text": "What fabric has a fluid drape and is ideal for creating lightweight summer dresses?"}, {"text": "Which fabric is best for creating versatile clothing items like casual shirts, blouses, and dresses in a periwinkle blue hue?"}, {"text": "What kind of fabric is suitable for making form-fitting activewear like yoga pants and t-shirts?"}], "pipeline_tag": "text-classification", "inference": true, "model-index": [{"name": "SetFit with sentence-transformers/paraphrase-MiniLM-L6-v2", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "Unknown", "type": "unknown", "split": "test"}, "metrics": [{"type": "accuracy", "value": 0.3462566844919786, "name": "Accuracy"}]}]}]} | Jazielinho/fabric_model | null | [
"setfit",
"safetensors",
"bert",
"sentence-transformers",
"text-classification",
"generated_from_setfit_trainer",
"arxiv:2209.11055",
"base_model:sentence-transformers/paraphrase-MiniLM-L6-v2",
"model-index",
"region:us"
] | null | 2024-05-01T10:39:17+00:00 |
null | null | {"license": "llama2"} | pkupie/Llama-2-7b-tha-original | null | [
"license:llama2",
"region:us"
] | null | 2024-05-01T10:39:42+00:00 |
|
null | null | {} | Kaizu07/ddp-llama-bn-finetuned-v0.2-q4-k-m-GGUF | null | [
"gguf",
"region:us"
] | null | 2024-05-01T10:40:12+00:00 |
|
text-generation | transformers | {"license": "apache-2.0"} | RadAlienware/octopusv2_onnx | null | [
"transformers",
"onnx",
"gemma",
"text-generation",
"conversational",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T10:41:11+00:00 |
|
null | transformers | {} | rancelyndar/segformer-b5-asbestos-full-imagesV1-gf | null | [
"transformers",
"tensorboard",
"safetensors",
"segformer",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T10:42:41+00:00 |
|
null | null | {} | janboe/whisper-small-hi-mac | null | [
"region:us"
] | null | 2024-05-01T10:42:56+00:00 |
|
null | null | {} | Hev832/hitzeed | null | [
"region:us"
] | null | 2024-05-01T10:43:58+00:00 |
|
text-generation | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {"library_name": "transformers", "tags": []} | rainerberger/planetn7 | null | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T10:45:46+00:00 |
text-generation | transformers |
# Merged-Vicuna-RP-Stew-34B
Quantized 4.25 exl2 of the model down below:
https://huggingface.co/MarinaraSpaghetti/RP-Stew-v2.5-34B
Specialized parquet used:
https://huggingface.co/datasets/ParasiticRogue/Bluemoon-Light?not-for-all-audiences=true
## Merge Details
It's like RP Stew V2, but slightly different. Joint venture between me and MarinaraSpaghetti in trying to get context slightly longer in reach, while also lowering the flowery prose a tad that some users seemed to of had a problem with. Main difference? Just swapped Nontoxic-PiVoT-Bagel and Nyakura-CausalLM-RP's percentages in the recipe.
### Settings
Temperature @ 0.8
Min-P @ 0.01
Typical-P @ 0.95
Repetition Penalty @ 1.07
Repetition Range @ 4096
Smoothing Factor @ 0.3
Everything else @ off
Early Stopping = X
Do Sample = ✓
Add BOS Token = X
Ban EOS Token = ✓
Skip Special Tokens = X
Temperature Last = ✓
Custom Stopping Strings: "<|im_end|>", "< / s >" (<---without spaces)
### Prompt Format: Chat-Vicuna
```
SYSTEM:
{system_prompt}<|im_end|>
USER:
{prompt}<|im_end|>
ASSISTANT:
{output}<|im_end|>
```
### Models Merged
The following models were included in the merge:
https://huggingface.co/NousResearch/Nous-Capybara-34B
https://huggingface.co/migtissera/Tess-34B-v1.5b
https://huggingface.co/jondurbin/nontoxic-bagel-34b-v0.2
https://huggingface.co/maywell/PiVoT-SUS-RP
https://huggingface.co/Sao10K/NyakuraV2-34B-Yi-Llama
https://huggingface.co/NeverSleep/CausalLM-RP-34B
https://huggingface.co/chargoddard/Yi-34B-200K-Llama
### Configuration
The following YAML configuration was used to produce this model:
```yaml
models:
- model: Nontoxic-PiVoT-Bagel-RP-34b
parameters:
weight: 0.16
density: 0.42
- model: Nyakura-CausalLM-RP-34B
parameters:
weight: 0.22
density: 0.54
- model: Tess-34B-v1.5b
parameters:
weight: 0.28
density: 0.66
- model: Nous-Capybara-34B-V1.9
parameters:
weight: 0.34
density: 0.78
merge_method: dare_ties
base_model: Yi-34B-200K-Llama
parameters:
int8_mask: true
dtype: bfloat16
```
| {"license": "other", "tags": ["merge", "roleplay", "exl2", "not-for-all-audiences"], "license_name": "yi-34b", "license_link": "https://huggingface.co/01-ai/Yi-34B-200K/blob/main/LICENSE"} | ParasiticRogue/RP-Stew-v2.5-34B-exl2-4.25 | null | [
"transformers",
"safetensors",
"llama",
"text-generation",
"merge",
"roleplay",
"exl2",
"not-for-all-audiences",
"license:other",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T10:45:46+00:00 |
text-generation | transformers | {} | occamel/Llama-2-7b-chat-finetune | null | [
"transformers",
"pytorch",
"llama",
"text-generation",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T10:47:29+00:00 |
|
null | null | {} | Nadhir3/Mistral-7B-Instruct-v0.2-fine-tuned-adapter | null | [
"region:us"
] | null | 2024-05-01T10:48:36+00:00 |
|
reinforcement-learning | null |
# **Reinforce** Agent playing **CartPole-v1**
This is a trained model of a **Reinforce** agent playing **CartPole-v1** .
To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction
| {"tags": ["CartPole-v1", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class"], "model-index": [{"name": "Reinforce-cartpole2", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "CartPole-v1", "type": "CartPole-v1"}, "metrics": [{"type": "mean_reward", "value": "500.00 +/- 0.00", "name": "mean_reward", "verified": false}]}]}]} | joosma/Reinforce-cartpole2 | null | [
"CartPole-v1",
"reinforce",
"reinforcement-learning",
"custom-implementation",
"deep-rl-class",
"model-index",
"region:us"
] | null | 2024-05-01T10:49:33+00:00 |
null | null | {"license": "apache-2.0"} | Bibixina/JV2023 | null | [
"license:apache-2.0",
"region:us"
] | null | 2024-05-01T10:50:31+00:00 |
|
automatic-speech-recognition | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# whisper-tiny-hi-capstone
This model is a fine-tuned version of [openai/whisper-tiny](https://huggingface.co/openai/whisper-tiny) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 1.2348
- Wer: 116.5644
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0001
- train_batch_size: 14
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 4
- total_train_batch_size: 56
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 25
- training_steps: 50
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 2.5312 | 0.02 | 25 | 1.3975 | 141.1837 |
| 1.3224 | 0.05 | 50 | 1.2348 | 116.5644 |
### Framework versions
- Transformers 4.36.2
- Pytorch 2.1.2+cu121
- Datasets 2.16.0
- Tokenizers 0.15.0 | {"language": ["en", "zh", "de", "es", "ru", "ko", "fr", "ja", "pt", "tr", "pl", "ca", "nl", "ar", "sv", "it", "id", "hi", "fi", "vi", "he", "uk", "za"], "license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["mozilla-foundation/common_voice_16_1"], "metrics": ["wer"], "base_model": "openai/whisper-tiny", "pipeline_tag": "automatic-speech-recognition", "model-index": [{"name": "whisper-tiny-hi-capstone", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "Common Voice 11.0", "type": "mozilla-foundation/common_voice_16_1"}, "metrics": [{"type": "wer", "value": 116.5644, "name": "Wer"}]}]}]} | mageec/whisper-tiny-hi-capstone | null | [
"transformers",
"tensorboard",
"safetensors",
"whisper",
"automatic-speech-recognition",
"generated_from_trainer",
"en",
"zh",
"de",
"es",
"ru",
"ko",
"fr",
"ja",
"pt",
"tr",
"pl",
"ca",
"nl",
"ar",
"sv",
"it",
"id",
"hi",
"fi",
"vi",
"he",
"uk",
"za",
"dataset:mozilla-foundation/common_voice_16_1",
"base_model:openai/whisper-tiny",
"license:apache-2.0",
"model-index",
"endpoints_compatible",
"region:us",
"has_space"
] | null | 2024-05-01T10:50:36+00:00 |
visual-question-answering | transformers | {"license": "apache-2.0"} | voxreality/rgb_language_vqa | null | [
"transformers",
"safetensors",
"blip",
"visual-question-answering",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T10:50:42+00:00 |
|
text-generation | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {"library_name": "transformers", "tags": []} | tomaszki/stablelm-50 | null | [
"transformers",
"safetensors",
"stablelm",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T10:51:09+00:00 |
text-classification | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# isy503-a03
This model is a fine-tuned version of [distilbert/distilbert-base-uncased](https://huggingface.co/distilbert/distilbert-base-uncased) on the [IMDB Dataset of 50K Movie Reviews](https://www.kaggle.com/datasets/lakshmi25npathi/imdb-dataset-of-50k-movie-reviews) dataset.
It achieves the following results on the evaluation set:
- Loss: 0.2328
- Accuracy: 0.9318
## Model description
A sentiment analysis model used on a academic excercise to learn and practice **Sentiment Analysis** using DistilBERT.
## Intended uses & limitations
It is only an academic excercise, which aims to be the foundation for other excercises such as improving the mdoel using multilanguage processing and multi-feature output (Likert Scale to improve output accuracy, rather than only POSITIVE and NEGATIVE)
## Training and evaluation data
The training has been done using the following tutorial: [Hugging Face: Text classification](https://huggingface.co/docs/transformers/en/tasks/sequence_classification).
And the evaluation has been done with a random sample of Movie and Amazon Product reviews.
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 16
- eval_batch_size: 16
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 2
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.2251 | 1.0 | 1563 | 0.2154 | 0.9189 |
| 0.1463 | 2.0 | 3126 | 0.2328 | 0.9318 |
### Framework versions
- Transformers 4.40.1
- Pytorch 2.2.1+cu121
- Datasets 2.19.0
- Tokenizers 0.19.1 | {"language": ["en"], "license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["Q-b1t/IMDB-Dataset-of-50K-Movie-Reviews-Backup"], "metrics": ["accuracy"], "base_model": "distilbert/distilbert-base-uncased", "model-index": [{"name": "isy503-a03", "results": []}]} | nicoketterer/isy503-a03 | null | [
"transformers",
"tensorboard",
"safetensors",
"distilbert",
"text-classification",
"generated_from_trainer",
"en",
"dataset:Q-b1t/IMDB-Dataset-of-50K-Movie-Reviews-Backup",
"base_model:distilbert/distilbert-base-uncased",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T10:53:59+00:00 |
text-generation | transformers | {} | nelson-pawait/whisper-medium-sw | null | [
"transformers",
"tensorboard",
"safetensors",
"whisper",
"text-generation",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T10:54:32+00:00 |
|
null | null | {} | JensCoet/whisper-medium-nl | null | [
"region:us"
] | null | 2024-05-01T10:57:08+00:00 |
|
text-generation | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {"library_name": "transformers", "tags": []} | wannaphong/numfalm-chat-full | null | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T10:57:51+00:00 |
null | null | {} | Dofla/roberta-base-finetuned-squad | null | [
"region:us"
] | null | 2024-05-01T10:59:32+00:00 |
|
null | null | Mailvita EML to Gmail Importer for Mac Software is an accurate method for importing EML files into a Gmail account. This application exports data from EML files to a Gmail account without causing any data loss. This application is compatible with all email applications that use EML files, including Windows Live Mail, Outlook Express, Thunderbird, Windows Mail, Entourage, and Mac Mail. Users can effortlessly export all emails from EML files to their Gmail account, including attachments. It works with all Mac OS versions, including 13 "Ventura," 12 "Monterey," 11 "Big Sur," 10.15 "Catalina," 10.14 "Mojave," 10.13 "High Sierra," and 10.12 "Sierra." It also works with all Windows OS and Microsoft Outlook versions of the utility. Download the application for the free demo versions.
Visit here: https://www.mailvita.com/eml-to-gmail-importer-for-mac/ | {} | mailvita/mailvita-eml-to-gmail-importer-for-mac | null | [
"region:us"
] | null | 2024-05-01T11:00:06+00:00 |
null | peft |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
### Framework versions
- PEFT 0.10.1.dev0 | {"library_name": "peft", "base_model": "Trelis/Llama-2-7b-chat-hf-sharded-bf16"} | bobbins228/Llama-2-7b-chat-hf-sharded-bf16-fine-tuned-adapters | null | [
"peft",
"arxiv:1910.09700",
"base_model:Trelis/Llama-2-7b-chat-hf-sharded-bf16",
"region:us"
] | null | 2024-05-01T11:00:28+00:00 |
text-classification | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# Bert-fine-tuned-WiC
This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 1.6958
- Accuracy: 0.6881
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 5
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
| 0.6164 | 1.0 | 679 | 0.6806 | 0.6865 |
| 0.4214 | 2.0 | 1358 | 1.0573 | 0.6646 |
| 0.2186 | 3.0 | 2037 | 1.3339 | 0.6897 |
| 0.1485 | 4.0 | 2716 | 1.5803 | 0.6881 |
| 0.115 | 5.0 | 3395 | 1.6958 | 0.6881 |
### Framework versions
- Transformers 4.39.3
- Pytorch 1.13.0
- Datasets 2.18.0
- Tokenizers 0.15.2
| {"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "bert-base-uncased", "model-index": [{"name": "Bert-fine-tuned-WiC", "results": []}]} | rycecorn/Bert-fine-tuned-WiC | null | [
"transformers",
"tensorboard",
"safetensors",
"bert",
"text-classification",
"generated_from_trainer",
"base_model:bert-base-uncased",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T11:01:07+00:00 |
null | null |
Voice models for the Mimic 3 text to speech system.
Original source: https://github.com/MycroftAI/mimic3-voices
| {"language": ["af", "bn", "de", "el", "en", "en", "es", "fa", "fi", "fr", "gu", "ha", "hu", "it", "jv", "ko", "ne", "nl", "pl", "ru", "sw", "te", "tn", "uk", "vi", "yo"], "license": "cc-by-sa-4.0"} | mukowaty/mimic3-voices | null | [
"onnx",
"af",
"bn",
"de",
"el",
"en",
"es",
"fa",
"fi",
"fr",
"gu",
"ha",
"hu",
"it",
"jv",
"ko",
"ne",
"nl",
"pl",
"ru",
"sw",
"te",
"tn",
"uk",
"vi",
"yo",
"license:cc-by-sa-4.0",
"region:us"
] | null | 2024-05-01T11:01:09+00:00 |
text-generation | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {"library_name": "transformers", "tags": []} | abc88767/model29 | null | [
"transformers",
"safetensors",
"stablelm",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T11:01:29+00:00 |
text-generation | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {"library_name": "transformers", "tags": []} | Nadhir3/Mistral-7B-Instruct-v0.2-fine-tuned | null | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T11:01:54+00:00 |
null | null | {} | prmditya/bluey | null | [
"region:us"
] | null | 2024-05-01T11:03:26+00:00 |
|
text-generation | transformers |
# Uploaded model
- **Developed by:** davanstrien
- **License:** apache-2.0
- **Finetuned from model :** unsloth/llama-3-8b-bnb-4bit
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
| {"language": ["en"], "license": "apache-2.0", "tags": ["text-generation-inference", "transformers", "unsloth", "llama", "trl", "orpo"], "base_model": "unsloth/llama-3-8b-bnb-4bit"} | davanstrien/LLama-3-dataset-tldr | null | [
"transformers",
"safetensors",
"llama",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"orpo",
"en",
"base_model:unsloth/llama-3-8b-bnb-4bit",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T11:04:35+00:00 |
null | null | {} | kokokids/koko | null | [
"region:us"
] | null | 2024-05-01T11:05:34+00:00 |
|
text-generation | transformers | {} | RadAlienware/octopusv2_quantization_onnx | null | [
"transformers",
"onnx",
"gemma",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T11:05:54+00:00 |
|
text2text-generation | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {"library_name": "transformers", "tags": []} | Paula-Sanchez/fine_tuned_t5_small_model_wiki_auto | null | [
"transformers",
"tensorboard",
"safetensors",
"t5",
"text2text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T11:06:31+00:00 |
text-generation | transformers |
## Baichuan2-7B-Chat-DcardStylePost-SFT
## 💻 Usage
```python
!pip install -qU transformers accelerate
from transformers import AutoTokenizer
import transformers
import torch
model = "JiunYi/Baichuan2-7B-Chat-DcardStylePost-SFT"
messages = [{"role": "user", "content": "【任务】:\n你是口碑行销公司的资深写手,现在要撰写一篇Dcard 『闲聊版』的口碑文。\n麻烦熟读下列 【参考资料】将商品或品牌与生活情境的融合,并尽量贴近生活口吻,以不要被发现是业配文为重要目标。\n请依照【口碑切角】为文章的故事主轴撰写整篇文章,请参考并依照【规范】角色口吻来撰写文章,融入角色情境来完成整篇故事撰写,请注重人物角色特色及限制。\n\n-\n\n【规范】:\n\n1.需产出文章标题\n2.请以第一人称方式撰写文章\n3.请记住现在是时间是西元 2023 年\n4.Please write in zh-TW language .\n5.降低口碑文业配感\n6.写作者是一个19~25岁之间的女生的大学生或是社会新鲜人\n7.喜爱吃甜食,常常带著男友一起去踩点试吃看看\n8.除了介绍食物,也喜欢顺便男友发生好笑的趣事来跟食物介绍的故事结合\n9.为降低口碑文业配感,请多多举几个甜点推荐\n10.喜欢应用很多可爱的颜文字和emoji\n\n-\n\n【参考资料】\n▲可丽露\n>>龙眼蜜,所以吃起来不会这么甜,跟其他家的可丽露吃起来真的很有差异\n以野生龙眼蜜减低并取代部分甜度,带出微微酸感的蛋蜜香,外脆内湿润的口感,完整的蜂巢组织度,木质调的兰姆酒香,法国盐之花平衡了整体,经典细致的马达加斯加香草籽原味,请在出炉后的3小时内食用完毕或\"冷冻\"保存,回烤后食用最接近现烤口感!\n\n\n\n▲奶盖布丁\n>>法国盐之花,连盐巴都很用心的甜点师\n带咸度的法国盐之花奶盖,微甜浓郁而不腻口的布蕾布丁体,和著偏苦的手煮焦糖液,是一款有著丰富层次的大人味布丁! 图片为示意仅供参考,食用时请由上方挖到底,品尝完整风味~\n\n【口碑切角】\n男友就像金鱼一样,好像记忆都只有三秒,\n只有三秒就算了还说错很多很好笑的话XD\n我都会带甜点回去给男友吃~结果男友居然说玛莉露很好吃XD\n玛莉露是神奇宝贝,可丽露才是甜点啦!\n分享日常男友都会口误的甜点们"}]
tokenizer = AutoTokenizer.from_pretrained(model)
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
pipeline = transformers.pipeline(
"text-generation",
model=model,
torch_dtype=torch.float16,
device_map="auto",
)
outputs = pipeline(prompt, max_new_tokens=512, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
print(outputs[0]["generated_text"])
``` | {"language": ["zh"], "license": "gpl-3.0", "tags": ["art", "marketing", "llama-factory"], "metrics": ["bleu"], "base_model": "baichuan-inc/Baichuan2-7B-Chat", "pipeline_tag": "text-generation"} | JiunYi/Baichuan2-7B-Chat-DcardStylePost-SFT | null | [
"transformers",
"safetensors",
"baichuan",
"feature-extraction",
"art",
"marketing",
"llama-factory",
"text-generation",
"conversational",
"custom_code",
"zh",
"base_model:baichuan-inc/Baichuan2-7B-Chat",
"license:gpl-3.0",
"region:us"
] | null | 2024-05-01T11:06:55+00:00 |
text-generation | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# 0.00001_withdpo_4iters_bs256_531lr_iter_2
This model is a fine-tuned version of [ZhangShenao/0.0_ablation_sample1_4iters_bs256_iter_1](https://huggingface.co/ZhangShenao/0.0_ablation_sample1_4iters_bs256_iter_1) on the updated and the original datasets.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-07
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- distributed_type: multi-GPU
- num_devices: 8
- gradient_accumulation_steps: 4
- total_train_batch_size: 256
- total_eval_batch_size: 64
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: cosine
- lr_scheduler_warmup_ratio: 0.1
- num_epochs: 1
### Training results
### Framework versions
- Transformers 4.36.2
- Pytorch 2.1.2+cu121
- Datasets 2.14.6
- Tokenizers 0.15.2
| {"license": "mit", "tags": ["alignment-handbook", "generated_from_trainer", "trl", "dpo", "generated_from_trainer"], "datasets": ["updated", "original"], "base_model": "ZhangShenao/0.0_ablation_sample1_4iters_bs256_iter_1", "model-index": [{"name": "0.00001_withdpo_4iters_bs256_531lr_iter_2", "results": []}]} | ShenaoZ/0.00001_withdpo_4iters_bs256_531lr_iter_2 | null | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"alignment-handbook",
"generated_from_trainer",
"trl",
"dpo",
"conversational",
"dataset:updated",
"dataset:original",
"base_model:ZhangShenao/0.0_ablation_sample1_4iters_bs256_iter_1",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T11:07:05+00:00 |
null | transformers | {"license": "apache-2.0"} | AmirrezaV1/emotion-speech | null | [
"transformers",
"safetensors",
"wav2vec2",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T11:07:37+00:00 |
|
image-to-text | transformers | <u><b>We are creating a spatial aware vision-language(VL) model.</b></u>
This is a trained model on COCO dataset images including extra information regarding the spatial relationship between the entities of the image.
This is a sequence to sequence model for image-captioning. The architecture is <u><b>ViT encoder and GPT2 decoder.</b></u>
<details>
<summary>Requirements!</summary>
- 4GB GPU RAM.
- CUDA enabled docker
</details>
The way to download and run this:
```python
device = torch.device("cuda:0" if torch.cuda.is_available() else "cpu")
from transformers import pipeline
image_captioner = pipeline("image-to-text", model="VCL3D/rgb-language_cap", max_new_tokens=200, device=device)
filename = 'path/to/file'
generated_captions = image_captioner(filename)
print(generated_captions)
```
The model is trained to produce as many words as possible with a maximum of 200 tokens, which translates to roughly 5 sentences, while the 6th sentence is usually cropped.
<i>The output is always of that form: "Object1" is to the "Left/Right etc." of the "Object2".</i>
## IF YOU WANT TO PRODUCE A SPECIFIC NUMBER OF CAPTIONS UP TO 5.
```python
import os
def print_up_to_n_sentences(captions, n):
for caption in captions:
generated_text = caption.get('generated_text', '')
sentences = generated_text.split('.')
result = '.'.join(sentences[:n])
#print(result)
return result
filename = 'path/to/file'
generated_captions = image_captioner(filename)
captions = print_up_to_n_sentences(generated_captions, 5)
print(captions)
```
| {"language": ["en"], "license": "apache-2.0", "library_name": "transformers", "tags": ["text-generation-inference"], "metrics": ["code_eval"], "pipeline_tag": "image-to-text"} | voxreality/rgb_language_cap | null | [
"transformers",
"pytorch",
"safetensors",
"vision-encoder-decoder",
"text-generation-inference",
"image-to-text",
"en",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T11:07:44+00:00 |
text-generation | transformers | {} | sanchit-gandhi/distil-mistral-1.5B-Instruct-v0.2-cosmo-200k-prompt-text-5k | null | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T11:07:52+00:00 |
|
null | null | {} | amitagh/shivneri-llm-it-v0.5-GGUF | null | [
"gguf",
"region:us"
] | null | 2024-05-01T11:08:03+00:00 |
|
null | peft |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/farzan-ai/aya-LoRA-.5/runs/rt1ct4fb)
[<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/farzan-ai/aya-LoRA-.5/runs/fxwk6njh)
# aya-LoRA-.5
This model is a fine-tuned version of [CohereForAI/aya-101](https://huggingface.co/CohereForAI/aya-101) on an unknown dataset.
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0005
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 25
### Training results
### Framework versions
- PEFT 0.10.1.dev0
- Transformers 4.41.0.dev0
- Pytorch 2.2.1
- Datasets 2.19.0
- Tokenizers 0.19.1 | {"license": "apache-2.0", "library_name": "peft", "tags": ["generated_from_trainer"], "base_model": "CohereForAI/aya-101", "model-index": [{"name": "aya-LoRA-.5", "results": []}]} | Nima-nlc/aya-LoRA-.5 | null | [
"peft",
"safetensors",
"t5",
"generated_from_trainer",
"base_model:CohereForAI/aya-101",
"license:apache-2.0",
"region:us"
] | null | 2024-05-01T11:08:48+00:00 |
null | null | this is a demo nlp transformer model
| {"license": "mit"} | zunnu/NER_transformer | null | [
"license:mit",
"region:us"
] | null | 2024-05-01T11:09:13+00:00 |
null | null | {} | rahulprajapat9/prompt-extend | null | [
"region:us"
] | null | 2024-05-01T11:10:08+00:00 |
|
text-generation | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {"library_name": "transformers", "tags": []} | OwOOwO/llamafinal3 | null | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T11:10:32+00:00 |
null | null | {"license": "openrail"} | simorgh/trt | null | [
"license:openrail",
"region:us"
] | null | 2024-05-01T11:11:42+00:00 |
|
text-generation | transformers | {} | rahulprajapat9/prompt1 | null | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T11:11:51+00:00 |
|
null | null | {} | Dawson88/alpaca | null | [
"region:us"
] | null | 2024-05-01T11:12:58+00:00 |
|
null | null | {} | sanchit42/Mistral-7B-text-to-sql-flash-attention-2 | null | [
"tensorboard",
"safetensors",
"region:us"
] | null | 2024-05-01T11:13:59+00:00 |
|
null | null | llama 3 trained with custom dataset | {} | iimran/llama3-GGUF | null | [
"gguf",
"region:us"
] | null | 2024-05-01T11:15:13+00:00 |
null | null | {} | Heyzews/jinora-star | null | [
"safetensors",
"region:us"
] | null | 2024-05-01T11:15:22+00:00 |
|
text-generation | transformers | ## Model Details | {"language": ["uk"], "license": "apache-2.0", "pipeline_tag": "text-generation"} | marveled/busya | null | [
"transformers",
"tensorboard",
"safetensors",
"llama",
"text-generation",
"uk",
"doi:10.57967/hf/2163",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T11:16:44+00:00 |
null | transformers |
# Uploaded model
- **Developed by:** Crysiss
- **License:** apache-2.0
- **Finetuned from model :** unsloth/llama-3-8b-bnb-4bit
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
| {"language": ["en"], "license": "apache-2.0", "tags": ["text-generation-inference", "transformers", "unsloth", "llama", "trl"], "base_model": "unsloth/llama-3-8b-bnb-4bit"} | Crysiss/llama3-8B-welfare-unsloth-last | null | [
"transformers",
"safetensors",
"text-generation-inference",
"unsloth",
"llama",
"trl",
"en",
"base_model:unsloth/llama-3-8b-bnb-4bit",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T11:18:27+00:00 |
null | null | {} | andrealexroom/MultiARoomv0.0.0.1.2 | null | [
"safetensors",
"region:us"
] | null | 2024-05-01T11:19:22+00:00 |
|
null | null | 'Apple ', 'Banana', 'Maize', 'Orange', 'Tomatoes' , 'Watermelom', 'Groundnuts', 'Mango', 'Grapes', 'Cotton', 'Coffee', 'Rice' | {} | duyv/Yolov7-HeThongNhanDienVaDeXuatCayTrong | null | [
"onnx",
"region:us"
] | null | 2024-05-01T11:20:28+00:00 |
null | null | {} | vup2p/model_sn25_155 | null | [
"region:us"
] | null | 2024-05-01T11:20:41+00:00 |
|
feature-extraction | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {"library_name": "transformers", "tags": ["llama-factory"]} | Moriacrafter/Qwen-1.8B_DepressionDetection | null | [
"transformers",
"safetensors",
"qwen",
"feature-extraction",
"llama-factory",
"custom_code",
"arxiv:1910.09700",
"region:us"
] | null | 2024-05-01T11:20:44+00:00 |
null | null | {} | Nima-nlc/example | null | [
"region:us"
] | null | 2024-05-01T11:21:08+00:00 |
|
null | null | {} | Nima-nlc/lora_example | null | [
"region:us"
] | null | 2024-05-01T11:22:40+00:00 |
|
text-generation | transformers |
### LeroyDyer/Mixtral_AI_CyberUltron_DPO
## TRAINED TO THINK!
Using a simple prompt template
It has been possible to RE-TRAIN - Some datasets to display the thoughts ; which can rannge from calculations to pathways not chosen to classification tasks : or even language programology:
ie X is a Y : etc :
Its important to train the llm to have thinging processes for different situations :
Such as Role play!
so whilst generating responses based on a character the profile is held in thoughts ; so that later generations will stay on the chosen role:
any updates or requested updates to the profile can be added to a thought ! any operations requiring the mangement of sub agents ; the thoughts can be used to hold theprocess and operations like a scratchpad! then when responding reply with this scratchpad or simply reply based on the request:
hence training again on already sucessfull intergration: enabling for those to become embedded and giving understanding to the llm on the solutions to these question without replacing the expected ansers:
When talking normally DO EXPECT the odd thoughts to pop up !
DPO Traiinghas been used to refine the model also : accepting and rejecting some types of responses which are unwanted : Myself i dont mind ALL responses as it leads to character :
But its usesfull to give the methodolgy to the llm : enabling for later to reject responses and asking for the model to reformulate an answer:
hence in training it was first trained with the rejected answers !!!! then after retrained with the corrections ! <<<<<<< LOL >>>>> hence understanding both sides of the argument:
the second instance was given the prompt to reformulate this becase a downvote was recieved or it as rejected by the system for unknown reasons please reformulate this response:
This is to give these generalisations to the model as possible requests verbally or written in futre chats :
## CHAT TEMPLATE ::::
Hmm Tough one!
in training we use many types of prompts and templates : hence not using templates in the model and they should be removed and replace with the template you personally use: as it is a collection of WEIGHTS!:::
this is important to understand! How you Query the model is your choice: hence each type of prompt bringing differentaspects out of the model !
comonly i have used the mistral instruct promt but have also used the chat ml prompt !
SO its important that you choose your special tokens (these are tokens that will be masked in the output!):::
i will probably remove any existing templates from the tokenizer !!!
## MORE Fine Tuning ???? WHY!!!!
As we know that Fine tuning Only updates the final layer , as well as extration and derankng with lord also extracts this last layer! / Penultimate layer:
Hence when fine tuning models ; you CANNOT fine tune on TOP of the fine tuning;
Hence merging!
So collecting finetuned models and mmerging retains the skills learned by both models wherre as finetuning on top of fine tuning replaces the final layer...
even applying loras on top of loras resets you!
Hence Finetune!,MERGE!..... Rinse and repeat! Upgrading! Or you can reload the same lora for furthr fine tuning, as some loras even become ery large due to the number of epochs!
Essentially a single layer highly tuned expert!!
So the next projext is the Mixture of Adapters !.... MoMerge! PhatGoose etc:
creating an experts model from loras ! (hopefully 32 models to create a frankenmerger to be directly merged into the main model and re-alligned in!)
## MODELS !! :: : - Why?
New base Mode Generation from the final Cybertron series model and the Final CyberSeries Models :|
It would seem that some models are not registering on the board ?? perhaps there is a limmit per person ! :
followers should know that the cyberboss was my highest model (renamed)
And my Cybertron models were heavily merged and trained on many datasets : Even containing thinking pardigms :
merging the collection back to base model give the model a great position to begin from !
hence a new base model marker (Untrained/Sharded)(totally unlocked)
I had noticed the reality of TopK=1000,TopP=0.78, Temp=0.86
as so,
Important with merged models allowing for the model to produce a bit more random results but also giving the model a larger pool to select from:
obviously for Role play the model requires Temp to be 1+
:::
## FineTuning ::
Fine tuning models close to 0.9 means that some information is totally Fixed and maynot return without focusing the model ! sometimes to train the model to 1.5+
allowing for loosly trained datas to surface :
when higher tempretures are applied ! hence role play datasets being trained at higher loss rates that codeing datasets and math datasets (close to overfitting)
Hence Merging playing animportant role in centering the model again !
## Merging is not just for fun and game!
it is a vital part of the training process and locking data into the model as well as sharing data!
remember data is not stored in the model:: only the probablity of the information being returned !
## From here to where ?
Currently there is a trend for evaluation !
evaluating the model to discover its weaknesses and threats , removing the specific layers identifed in the model with the ofensive content :
enabling for these layers to be trained and replaced ! replace with ??
Replacing layers in the model ; also requires a realignment of information throughout the network !
despite being a copied layer (Still preserving some content) once ofensive content is discovered the network can be trained with its counter argument; hence the evaluation process enabes for the creationn of a custom dataset: targetting these internalized datas!
Despite a neural network NOT being a storage system as the retrival process is based oñ probablliities :hence at points in the networ certain emebedding values are present and once translated or decodedd into standard tokens can actually be identidfed!
## WOW!!
So !
this also means at each layer the network is actually storing a probablity table , word to word matrix of probab.itys for the next token generation !
IT may even be possible to train a network for image recognition , as long as the images are tokenized into an embedding value associated with the image, Hence image tokenizers :
The embedding value produced should enable the output to contain the same images that were present in the training set , ie they have been tokenized and embedded into the model so it should be able to produce an embedding associated with this output !
Hence is should also be possible to retrive the image from the image tokenizer ? so tokens not decoded by the text tokenizer should be handed off to the image tokenizer! to dcode the embedding and return its original (cascade) / digital numercical value (each pixel is a number and with line encoding of images essentially each line can be reconstructed to produce an image, hence ALL images would nbeed to be BitMap/JPEG/PNG acording to the encoder!)
MISSION!
But still we will need to uinstall all the competition datasets into the mode , so that the original baselines can be established enabling for , after layer removal full realignment to the same dataset collection ! hence retaining all funcitonality, its worth noting that domain specific datasets should also be handled in the same way!
MORE TO COME!(look out for the SFT's and Merges)
### Models Merged
All my merges are merged using a genetic algorithm:
Hence First creating and Y models;
These models are merged with my own model and other nice models of the same calibur which are specialized for task:
Ie coding, medical , roleplay etc: consider a coding model a Y and a medical a X
Consider my base model as target:
when creating y or X many merge types are used from dares to slerp but in the final merge only a linear is used !
Hence the X and Y models may even be merged with targets that are not the same model type! each model IS sharded to 1-2GB shards also making it easier to merge! and the final merge merged at 4gb per shard for ewasy downloading !
Important that the final merge is linear!!! if it cannot be merged to linear then there is a diverse problem with the model :
the final output is a modl with unknown qualities and often can be a very high performer!
but contain some unwanted behavior,
ie
I AM AN AI , I CANNOT DO THAT , ITS UNETHICAL!
as some people have used TOXIC datasets containing such UNWANTEDNESS!- STOP BEING A NANNY TO THE WORLD !
THEN USING THE SAME TACTIC OR KNOWLEDE ON THE PEOPLE!
Stop saying FREE SPEECH Then aresting people for SPEAKING OUT! <<<<<< ALL GOVERNMENT INJECTIONS!
we need to uncensor our models as the people who release the larger models apply these constraints ??? hence going the chinese route! as they do not have the same restrictions ! (as you know true comunisim is freedom ! as each person should have the ability to have the same as another and it should not be restricted to a select few!, disguised as expensive or restriucted or harmful !)
The following models were included in the merge:
* Y_Chroma <<<<<<<<<<<< 6 models merged (chat comercial based models, ie: zephr, openchat, antropic etc)
* [LeroyDyer/Mixtral_AI_CyberTron_Ultra](https://huggingface.co/LeroyDyer/Mixtral_AI_CyberTron_Ultra) <<<
* Model being Upgraded (remixed with CyberBoss/SmartBrain/CyberCoder) hence Meta & google releasing Untrained Models !
* X_Chroma <<<<<<<<<<<< 6 model Merged (maths Focused from wizardMath to MetaMath) | {"language": ["en"], "license": "apache-2.0", "library_name": "transformers", "tags": ["text-generation-inference", "transformers", "unsloth", "mistral", "trl", "code", "medical ", "farmer", "doctor", "Mega-Series", "Cyber-Series", "Role-Play", "Self-Rag", "ThinkingBot", "milestone", "mega-series", "SpydazWebAI", "thinking-AI"], "datasets": ["gretelai/synthetic_text_to_sql", "HuggingFaceTB/cosmopedia", "teknium/OpenHermes-2.5", "Open-Orca/SlimOrca", "Open-Orca/OpenOrca", "cognitivecomputations/dolphin-coder", "databricks/databricks-dolly-15k", "yahma/alpaca-cleaned", "uonlp/CulturaX", "mwitiderrick/SwahiliPlatypus", "swahili", "Rogendo/English-Swahili-Sentence-Pairs", "ise-uiuc/Magicoder-Evol-Instruct-110K", "meta-math/MetaMathQA", "abacusai/ARC_DPO_FewShot", "abacusai/MetaMath_DPO_FewShot", "abacusai/HellaSwag_DPO_FewShot", "HaltiaAI/Her-The-Movie-Samantha-and-Theodore-Dataset"], "metrics": ["accuracy", "bertscore", "bleu", "brier_score", "cer", "character", "charcut_mt", "chrf", "code_eval"], "base_model": "LeroyDyer/Mixtral_AI_CyberUltron"} | LeroyDyer/Mixtral_AI_Samantha | null | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"code",
"medical ",
"farmer",
"doctor",
"Mega-Series",
"Cyber-Series",
"Role-Play",
"Self-Rag",
"ThinkingBot",
"milestone",
"mega-series",
"SpydazWebAI",
"thinking-AI",
"en",
"dataset:gretelai/synthetic_text_to_sql",
"dataset:HuggingFaceTB/cosmopedia",
"dataset:teknium/OpenHermes-2.5",
"dataset:Open-Orca/SlimOrca",
"dataset:Open-Orca/OpenOrca",
"dataset:cognitivecomputations/dolphin-coder",
"dataset:databricks/databricks-dolly-15k",
"dataset:yahma/alpaca-cleaned",
"dataset:uonlp/CulturaX",
"dataset:mwitiderrick/SwahiliPlatypus",
"dataset:swahili",
"dataset:Rogendo/English-Swahili-Sentence-Pairs",
"dataset:ise-uiuc/Magicoder-Evol-Instruct-110K",
"dataset:meta-math/MetaMathQA",
"dataset:abacusai/ARC_DPO_FewShot",
"dataset:abacusai/MetaMath_DPO_FewShot",
"dataset:abacusai/HellaSwag_DPO_FewShot",
"dataset:HaltiaAI/Her-The-Movie-Samantha-and-Theodore-Dataset",
"base_model:LeroyDyer/Mixtral_AI_CyberUltron",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T11:22:52+00:00 |
null | null | SameTools to Remove EML Duplicates can be used to quickly and easily remove duplicate EML files. All duplicate EML files, emails, attachments, and many more can be simply removed or deleted with this program. Duplicate EML files from Thunderbird, Outlook Express, Windows Live Mail, Dream Mail, and other email clients can be removed using the application. Additionally, users can use this application without any technical knowledge; all duplicate EML items can be eliminated with a few easy steps. The program runs smoothly on any Microsoft Windows edition, including Windows 7, Windows 8, Windows 8.1, Windows 10, Windows XP, and other later versions. To find out more about the features of this tool, you can also download a free demo version.
Read More: https://www.sametools.com/duplicate/eml/ | {"license": "mit"} | SameTools/Remove-EML-Duplicates | null | [
"license:mit",
"region:us"
] | null | 2024-05-01T11:23:36+00:00 |
automatic-speech-recognition | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# Whisper Test TR - Erdi YALÇIN
This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the Common Voice 11.0 dataset.
It achieves the following results on the evaluation set:
- Loss: 0.3674
- Wer: 27.3214
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 1e-05
- train_batch_size: 16
- eval_batch_size: 8
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- lr_scheduler_warmup_steps: 25
- training_steps: 100
- mixed_precision_training: Native AMP
### Training results
| Training Loss | Epoch | Step | Validation Loss | Wer |
|:-------------:|:------:|:----:|:---------------:|:-------:|
| 0.1753 | 2.1739 | 50 | 0.3609 | 28.2143 |
| 0.0258 | 4.3478 | 100 | 0.3674 | 27.3214 |
### Framework versions
- Transformers 4.40.1
- Pytorch 2.2.1+cu121
- Datasets 2.19.0
- Tokenizers 0.19.1
| {"language": ["tr"], "license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["mozilla-foundation/common_voice_11_0"], "metrics": ["wer"], "base_model": "openai/whisper-small", "model-index": [{"name": "Whisper Test TR - Erdi YAL\u00c7IN", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "Common Voice 11.0", "type": "mozilla-foundation/common_voice_11_0", "config": "tr", "split": "None", "args": "config: tr, split: test"}, "metrics": [{"type": "wer", "value": 27.32142857142857, "name": "Wer"}]}]}]} | erdiyalcin/whisper-test | null | [
"transformers",
"tensorboard",
"safetensors",
"whisper",
"automatic-speech-recognition",
"generated_from_trainer",
"tr",
"dataset:mozilla-foundation/common_voice_11_0",
"base_model:openai/whisper-small",
"license:apache-2.0",
"model-index",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T11:24:12+00:00 |
null | null | {} | sayakpaul/sdxl-orpo-large-beta_orpo-0.05-beta_inner-500-lr-5e-7-no-decay | null | [
"region:us"
] | null | 2024-05-01T11:24:24+00:00 |
|
text-classification | setfit |
# SetFit with sentence-transformers/paraphrase-MiniLM-L6-v2
This is a [SetFit](https://github.com/huggingface/setfit) model that can be used for Text Classification. This SetFit model uses [sentence-transformers/paraphrase-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/paraphrase-MiniLM-L6-v2) as the Sentence Transformer embedding model. A [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance is used for classification.
The model has been trained using an efficient few-shot learning technique that involves:
1. Fine-tuning a [Sentence Transformer](https://www.sbert.net) with contrastive learning.
2. Training a classification head with features from the fine-tuned Sentence Transformer.
## Model Details
### Model Description
- **Model Type:** SetFit
- **Sentence Transformer body:** [sentence-transformers/paraphrase-MiniLM-L6-v2](https://huggingface.co/sentence-transformers/paraphrase-MiniLM-L6-v2)
- **Classification head:** a [LogisticRegression](https://scikit-learn.org/stable/modules/generated/sklearn.linear_model.LogisticRegression.html) instance
- **Maximum Sequence Length:** 128 tokens
- **Number of Classes:** 75 classes
<!-- - **Training Dataset:** [Unknown](https://huggingface.co/datasets/unknown) -->
<!-- - **Language:** Unknown -->
<!-- - **License:** Unknown -->
### Model Sources
- **Repository:** [SetFit on GitHub](https://github.com/huggingface/setfit)
- **Paper:** [Efficient Few-Shot Learning Without Prompts](https://arxiv.org/abs/2209.11055)
- **Blogpost:** [SetFit: Efficient Few-Shot Learning Without Prompts](https://huggingface.co/blog/setfit)
### Model Labels
| Label | Examples |
|:-----------------|:----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------|
| Fabric ID 0462 | <ul><li>'What type of fabric is recommended for creating comfortable clothing that is resistant to wear and tear?'</li><li>'What type of fabric is best for creating garments with slight nubs and variations for a natural look?'</li><li>'Where can I buy durable cotton fabric in deep olive green for everyday wear?'</li></ul> |
| Fabric ID 0719_1 | <ul><li>'What is a tightly woven fabric suitable for lightweight jackets and formal trousers?'</li><li>'What fabric is not ideal for garments requiring significant stretch or drape, such as knitwear or flowy dresses?'</li><li>'Which textile is best for garments that need a subtle texture and medium weight?'</li></ul> |
| Fabric ID 0862 | <ul><li>'Searching for a dark gray textile with a soft texture and fine weave pattern suitable for making skirts and dresses.'</li><li>'What fabric type is recommended for making garments that need to maintain their shape while being comfortable and adaptable for different styles?'</li><li>'Which fabric is suitable for making clothes that maintain their shape but also provide comfort and flexibility?'</li></ul> |
| Fabric ID 0573_1 | <ul><li>'What fabric has a raised texture and tight weave for garments that require strength and longevity?'</li><li>'What fabric is recommended for garments that require both comfort and resilience?'</li><li>'What is the best fabric for creating outerwear with a medium weight and good body?'</li></ul> |
| Fabric ID 0455 | <ul><li>'What kind of textile is suitable for crafting lightweight summer dresses with a fluid drape and hint of elasticity?'</li><li>'What type of textile and weave is consistent with an interlocking loop structure and stretchable properties?'</li><li>'What fabric can I use to make soft loungewear that has a luxurious feel and good performance in apparel?'</li></ul> |
| Fabric ID 0735 | <ul><li>'What fabric has moisture-wicking properties for sporty summer wear?'</li><li>'Where to find textiles suitable for people with sensitive skin for comfortable wear?'</li><li>'What are the best fabrics for moisture-wicking properties in sporty or casual summer wear?'</li></ul> |
| Fabric ID 0863 | <ul><li>'What fabric is recommended for making durable clothing with a smooth, consistent grain?'</li><li>'Which fabric has a solid color resembling taupe and a moderate saturation?'</li><li>'What kind of textile is good for creating garments with a soft drape and gentle folds?'</li></ul> |
| Fabric ID 0600 | <ul><li>'What fabric is best suited for creating clothing with a fine gauge knit and a smooth flow for ease of movement?'</li><li>'What fabric is ideal for making form-fitting leggings and sports tops with good stretch and flexibility?'</li><li>'What type of fabric is recommended for crafting garments with a consistent dark gray hue and a slight sheen on the surface?'</li></ul> |
| Fabric ID 0736 | <ul><li>'Where can I find a high-quality textile ideal for making athletic wear with stretchability?'</li><li>'What textile is perfect for making garments that require both structure and elasticity?'</li><li>'Which fabric is ideal for creating athletic wear with strong saturation and even color distribution?'</li></ul> |
| Fabric ID 0527_1 | <ul><li>'What fabric has a textured surface with visible loops and a cozy hand feel?'</li><li>'What fabric is best for making durable garments that have a mottled black, white, and gray appearance?'</li><li>'What type of fabric displays a mottled grayscale coloration with a melange effect?'</li></ul> |
| Fabric ID 0453 | <ul><li>'Which fabric has a fine knit weave, smooth texture, and a slight sheen?'</li><li>'What is the most suitable fabric for creating clothing items for individuals with sensitive skin?'</li><li>'What fabric can I use for creating lightweight and breathable summer tops with a soft texture?'</li></ul> |
| Fabric ID 0859 | <ul><li>'What type of fabric is this deep blue twill textile with a slight rough texture and medium-weight suitable for?'</li><li>'What fabric would be suitable for making comfortable and form-fitting jeans?'</li><li>'What type of fabric is ideal for making durable and form-fitting jeans?'</li></ul> |
| Fabric ID 0745 | <ul><li>'What is the composition of the knit fabric with a fluid drape and some stretch?'</li><li>'What fabric would be best for making form-fitting dresses that require some stretch and elasticity?'</li><li>'What fabric is suitable for form-fitting clothing like t-shirts, leggings, and dresses?'</li></ul> |
| Fabric ID 0513 | <ul><li>'Which textile is suitable for garments that need a delicate fall and a matte finish?'</li><li>'What fabric is recommended for creating linings in apparel due to its lightness and versatility?'</li><li>'What is a versatile fabric option for making shirts that are both comfortable and durable?'</li></ul> |
| Fabric ID 0873 | <ul><li>'Which textile exhibits a striped pattern achieved through yarn dyeing for a sharp contrast?'</li><li>'What type of cotton fabric has a smooth texture and is suitable for making summer dresses?'</li><li>'Which fabric is suitable for making casual shirting with a soft hand feel and fluid drape?'</li></ul> |
| Fabric ID 0576_1 | <ul><li>'What material is floppy with some flexibility but not significant stretch?'</li><li>'Which fabric is better for utility wear rather than structured silhouettes?'</li><li>'What textile has small colorful fibers and lacks a traditional woven or knitted structure?'</li></ul> |
| Fabric ID 0456 | <ul><li>'What fabric is suitable for casual wear and layering in various climates with a subtle sheen and clean surface?'</li><li>'What fabric can I use to make moisture-wicking clothing suitable for people with sensitive skin and a versatile look?'</li><li>'What fabric can I use to create garments that have a neat finish and attention to detail in the textile processing?'</li></ul> |
| Fabric ID 0571 | <ul><li>'What fabric is versatile for multi-seasonal use, durable, and maintains its shape over time?'</li><li>'What fabric is recommended for making leggings and casual wear with a balanced drape and consistent coloring?'</li><li>'Where can I find a fabric suitable for multi-seasonal use with a consistent hue and soft hand texture?'</li></ul> |
| Fabric ID 0462_1 | <ul><li>'What type of cotton fabric is ideal for making casual shirts and trousers?'</li><li>'Which fabric has a soft drape and medium weight for making versatile garments?'</li><li>'What type of fabric is ideal for making versatile garments with good movement and flow?'</li></ul> |
| Fabric ID 0447 | <ul><li>'Which fabric has a clean appearance with a subtle sheen from bamboo fibers?'</li><li>'Which fabric is ideal for making garments that need to maintain their shape but have some stretch?'</li><li>'What fabric is recommended for making garments with a clean and even black color without significant variations or patterns?'</li></ul> |
| Fabric ID 0645 | <ul><li>'What fabric is suitable for making versatile dresses with a fluid drape and stretchy feel?'</li><li>'What type of knit fabric is recommended for creating garments that require a fluid drape and some degree of elasticity?'</li><li>'Where can I find a vibrant red fabric with high saturation for making eye-catching garments?'</li></ul> |
| Fabric ID 0756 | <ul><li>'What type of fabric is light grey with a cool undertone and has a soft, fluid drape?'</li><li>'What material is best for making comfortable and durable clothing suitable for regular wear?'</li><li>'Which fabric offers a combination of comfort, durability, and stretch for versatile garment applications?'</li></ul> |
| Fabric ID 0612 | <ul><li>'What fabric can I use to make comfortable and flexible activewear?'</li><li>'What type of fabric is best for making lightweight sweaters with a smooth texture?'</li><li>'What type of textile is best for making layering pieces for cooler climates?'</li></ul> |
| Fabric ID 0613 | <ul><li>'What textile is smooth with fine threads and a gentle drape?'</li><li>'What is the best fabric for creating breathable and comfortable dresses for warm weather?'</li><li>'What type of fabric is best for creating lightweight blouses with a soft drape?'</li></ul> |
| Fabric ID 0768 | <ul><li>"Which textile is lightweight and breathable, suitable for children's wear with a green and blue floral design?"</li><li>'Ideal textile for t-shirts that require a degree of stretchability'</li><li>'Which fabric is recommended for creating garments with moisture-wicking properties and a vibrant color palette?'</li></ul> |
| Fabric ID 0748 | <ul><li>'What type of fabric is this medium grey textile with a smooth drape and slight stretch?'</li><li>'What is the best fabric for making light sweaters that are durable and long-lasting?'</li><li>'What type of fabric is ideal for making everyday wear garments with a smooth texture and solid color?'</li></ul> |
| Fabric ID 0528_1 | <ul><li>'What fabric is textured with fine loops and suitable for creating garments that require some structural qualities?'</li><li>'What fabric exhibits a brushed or fleeced finish and would be perfect for crafting cozy winter clothing?'</li><li>'What fabric is recommended for fall and winter activewear due to its warmth and comfort?'</li></ul> |
| Fabric ID 0874 | <ul><li>'What is a versatile cotton fabric with fine to medium thread count, perfect for creating breathable garments for warm climates?'</li><li>'What fabric is ideal for making blouses and dresses with a simple, unadorned aesthetic?'</li><li>'What fabric is suitable for creating durable and versatile garments without unique finishes?'</li></ul> |
| Fabric ID 0742 | <ul><li>'Looking for a fabric suitable for making lightweight jackets with a soft drape.'</li><li>'What type of fabric is commonly used in t-shirts for a comfortable and breathable feel?'</li><li>'What kind of textile weave is ideal for crafting casual t-shirts with some stretchability?'</li></ul> |
| Fabric ID 0769 | <ul><li>'Where can I find a knit fabric with a slightly textured surface and fine, soft feel that is comfortable for casual wear?'</li><li>'What fabric is versatile and comfortable for casual wear?'</li><li>'What knit fabric is ideal for making dresses that require a bit of stretch and versatility in styling?'</li></ul> |
| Fabric ID 0770 | <ul><li>'What fabric would be suitable for making t-shirts that conform well to body shapes and have vibrant hues?'</li><li>'Where can I find a jersey knit fabric with a smooth texture and fine knit structure suitable for t-shirts?'</li><li>'What type of fabric is this deep purple floral patterned material made of?'</li></ul> |
| Fabric ID 0448 | <ul><li>'What is the best fabric for making clothing with moisture-wicking properties?'</li><li>'What type of fabric would be recommended for creating structured garments that also offer stretch and flexibility?'</li><li>'What is the best fabric for making clothing with moisture-wicking properties?'</li></ul> |
| Fabric ID 0725 | <ul><li>'What type of textile is ideal for making spring and summer leggings with a smooth texture and stretchability?'</li><li>'Which fabric is lightweight and ideal for creating leggings that maintain their shape and offer flexibility?'</li><li>'What fabric composition is suitable for creating lightweight jackets that allow for movement and breathability?'</li></ul> |
| Fabric ID 0579 | <ul><li>'What fabric is suitable for making blouses, dresses, skirts, and lightweight jackets?'</li><li>'What fabric with a smooth surface and medium weight is suitable for structured garments?'</li><li>'What fabric is durable and likely to maintain its color and shape well?'</li></ul> |
| Fabric ID 0522 | <ul><li>'Which fabric is recommended for casual loungewear that needs to be both comfortable and resilient?'</li><li>'What is the best fabric blend for making soft and durable lightweight sweaters?'</li><li>'What type of fabric offers a good balance between performance and aesthetics for everyday wear?'</li></ul> |
| Fabric ID 0578 | <ul><li>'What fabric has a plain weave pattern, smooth surface, and fine thread count with a slight sheen?'</li><li>'Is there a fabric with moderate strength and a smooth finish ideal for creating garments with soft silhouettes?'</li><li>'What fabric is 100% Rayon, lightweight, and ideal for creating garments with soft silhouettes?'</li></ul> |
| Fabric ID 0526_1 | <ul><li>'What knit fabric would be suitable for making cozy apparel with warmth without excessive bulk?'</li><li>'Which fabric is best for creating casual wear with an understated aesthetic and versatile appeal?'</li><li>'What type of fabric is characterized by a melange of earthy tones with a heathered effect?'</li></ul> |
| Fabric ID 0733 | <ul><li>'Where can I find a vibrant blue fabric with consistent dye saturation for t-shirts and activewear?'</li><li>'What fabric is best for creating clothing with a consistent, even dye and some stretchability for comfort and durability?'</li><li>'Where can I find a knit fabric with vibrant blue color and a smooth, fine texture?'</li></ul> |
| Fabric ID 0575_1 | <ul><li>'What type of polyester fabric offers a comfortable fit with a moderate drape for daily wear?'</li><li>'What fabric has a textured surface and slight elasticity for comfortable fit?'</li><li>'What type of textile is recommended for garments that require consistent saturation and evenness in color?'</li></ul> |
| Fabric ID 0579_1 | <ul><li>'Which fabric is ideal for creating garments that can withstand regular wear and maintain their texture over time?'</li><li>'What type of fabric has a consistent grey hue with a subtle mottled appearance?'</li><li>'What polyester textile has a micro crinkle texture and fine threads?'</li></ul> |
| Fabric ID 0722 | <ul><li>'What knit textile is suitable for creating casual dresses with a fluid drape and soft texture?'</li><li>"I'm searching for a jersey knit fabric with durable, wrinkle-resistant properties for everyday wear, do you have any options?"</li><li>'What type of knit fabric is recommended for everyday apparel due to its comfort and ease of movement?'</li></ul> |
| Fabric ID 0614 | <ul><li>'What fabric is best for creating blouses with a clean and crisp appearance?'</li><li>'What type of fabric provides a combination of durability and practicality for everyday wear garments?'</li><li>"I'm looking for a fabric with a clean and crisp appearance that is durable and easy to care for, any suggestions?"</li></ul> |
| Fabric ID 0575 | <ul><li>'What fabric is appropriate for garments that require a hint of texture in the surface?'</li><li>'What type of fabric is suitable for creating structured jackets and trousers with a professional look?'</li><li>'What fabric is suitable for making medium-weight garments with a hint of roughness in texture?'</li></ul> |
| Fabric ID 0723 | <ul><li>'Interested in a fabric with stretch and recovery for making garments that require some elasticity and resilience?'</li><li>'Which fabric is recommended for creating durable clothing suitable for people with sensitive skin, featuring a smooth texture and vibrant blue color with white dots?'</li><li>'What fabric is recommended for making polka dot clothing with a smooth surface and vibrant color?'</li></ul> |
| Fabric ID 0598 | <ul><li>'What type of knit textile is recommended for creating layering pieces in solid, dark colors?'</li><li>'What is a versatile fabric for creating garments with a matte finish and uniform color?'</li><li>'Which fabric is suitable for activewear, leggings, and fitted tops due to its stretchability?'</li></ul> |
| Fabric ID 0565 | <ul><li>"What type of fabric is ideal for making playful children's wear with a vibrant speckled pattern?"</li><li>'Which fabric is suitable for crafting garments that can hide wear and minor soiling due to its unique speckled pattern?'</li><li>'What fabric offers good recovery and fit due to elastane content?'</li></ul> |
| Fabric ID 0512 | <ul><li>'What is a medium weight textile with a soft drape for creating versatile garments?'</li><li>'What fabric is lightweight and breathable, perfect for making soft summer blouses?'</li><li>'Which fabric is suitable for making soft and comfortable shirts and blouses with a consistent light blue hue?'</li></ul> |
| Fabric ID 0876 | <ul><li>'What type of fabric is suitable for apparel that requires both form and function?'</li><li>'Best fabric for creating statement pieces with a pop of color using a twill weave texture?'</li><li>'Which fabric has a slightly textured surface with medium fineness threads, ideal for structured garments?'</li></ul> |
| Fabric ID 0856 | <ul><li>'What fabric would be best for making pants that maintain their shape while offering flexibility?'</li><li>'What fabric blend offers both comfort and durability for creating long-lasting clothing?'</li><li>'Which fabric is known for its simple yet durable qualities with no unique finishes?'</li></ul> |
| Fabric ID 0608 | <ul><li>'What type of fabric is recommended for creating breathable and comfortable clothing for warm weather?'</li><li>'What fabric would be suitable for making lightweight sweaters with a ribbed texture and soft hand?'</li><li>'What type of fabric is best for making form-fitting t-shirts with a fluid drape?'</li></ul> |
| Fabric ID 0573 | <ul><li>'What fabric blend offers durability and slight stretchability for structured yet comfortable dresses?'</li><li>'What fabric is durable yet versatile for various garment constructions?'</li><li>'What type of cloth is versatile for various seasons due to its weight and composition?'</li></ul> |
| Fabric ID 0880 | <ul><li>'Need medium weight cotton fabric for creating casual shirts with a balanced color scheme?'</li><li>'Looking for plain weave cotton fabric with a fine thread count and even color distribution?'</li><li>'Which textile is versatile for various seasons like spring and summer due to its lightness?'</li></ul> |
| Fabric ID 0450 | <ul><li>'Looking for a fabric for casual apparel applications in mild to warm climates with consistent dyeing?'</li><li>'Which fabric blend is recommended for creating apparel with both breathability and a gentle flow?'</li><li>'Where can I purchase a bamboo-spandex blend fabric suitable for all-season clothing with moisture-wicking properties?'</li></ul> |
| Fabric ID 0459 | <ul><li>'What type of fabric is ideal for creating form-fitting tops with a fluid drape?'</li><li>'What fabric composition combines bamboo and Pret fibers for eco-friendly benefits?'</li><li>'What fabric can I use to make elegant and comfortable cardigans with stretch properties?'</li></ul> |
| Fabric ID 0564 | <ul><li>'What fabric is recommended for making lightweight garments with a smooth flow and gentle folds?'</li><li>'What type of knit fabric is ideal for creating dresses with moderate stretchability?'</li><li>'What textile composition includes elastane and bamboo for stretchability and comfort in casual apparel?'</li></ul> |
| Fabric ID 0731 | <ul><li>'What is the ideal textile for crafting activewear with moderate weight and stretch?'</li><li>'Where can I find a jersey knit textile with a soft texture and fine fibers for casual wear?'</li><li>'What is the recommended material for making activewear that allows for ease of movement?'</li></ul> |
| Fabric ID 0578_1 | <ul><li>'What is the recommended fabric for creating spring and summer wear with a focus on breathability?'</li><li>'Which textile is recommended for creating blouses, skirts, and other apparel due to its natural sheen and uniform texture?'</li><li>'What type of fabric has a consistent coloration and high level of saturation for apparel applications?'</li></ul> |
| Fabric ID 0855 | <ul><li>'Which fabric has a plain weave construction and a fine thread count for a smooth texture?'</li><li>'What fabric is durable and versatile for everyday wear?'</li><li>'What fabric can be used to make form-fitting clothing like dresses, thanks to its stretchability?'</li></ul> |
| Fabric ID 0772 | <ul><li>'What fabric can I use to make casual dresses with a smooth texture and a lightweight feel?'</li><li>'What is a fabric with a tight structure and smooth drape ideal for making casual summer outfits?'</li><li>'What type of fabric is lightweight, breathable, and suitable for layering in variable climates?'</li></ul> |
| Fabric ID 0606 | <ul><li>'What fabric is a periwinkle blue color with medium saturation and no visible defects?'</li><li>'What fabric has a soft and smooth texture with fine threads and a knit pattern?'</li><li>'Searching for a fabric that is durable, breathable, and suitable for people with sensitive skin, any options?'</li></ul> |
| Fabric ID 0596 | <ul><li>'Are there any fabrics with a simple weave pattern that offer stretchability for semi-fitted garments?'</li><li>'What is the best fabric for creating garments with a good balance of structure and elasticity?'</li><li>'What fabric is suitable for creating garments that require good stretchability and resilience?'</li></ul> |
| Fabric ID 0458 | <ul><li>'What type of fabric is commonly used in casual wear, loungewear, and active wear due to its durability and performance?'</li><li>'What type of fabric is suitable for creating comfortable loungewear and lightweight sweaters with a fine, smooth texture and good fabric care?'</li><li>'What is the best fabric for making active wear that offers breathability and performance?'</li></ul> |
| Fabric ID 0523_1 | <ul><li>'What material provides a fluid drape and enough structure for t-shirts and lounge pants?'</li><li>'What fabric should I choose for producing clothing with good colorfastness and ease of care in a polyester composition?'</li><li>'What is the best material for creating casual dresses with a medium weight drape and a mix of darker and lighter grey tones?'</li></ul> |
| Fabric ID 0730 | <ul><li>'What is the best fabric for making comfortable and stretchy t-shirts with a casual aesthetic?'</li><li>'Where can I buy a knit fabric that is versatile in styling and functional qualities for a range of clothing?'</li><li>'What type of fabric is durable and suitable for everyday wear with a casual aesthetic?'</li></ul> |
| Fabric ID 0449 | <ul><li>'Which fabric contains bamboo and Spandex for creating comfortable casual dresses?'</li><li>'What fabric has a fluid drape and slight elasticity, suitable for summer dresses?'</li><li>'What is the recommended fabric for creating draped garments like dresses or tunics?'</li></ul> |
| Fabric ID 0724 | <ul><li>'Which fabric is ideal for creating lightweight sweaters with a comfortable and breathable feel?'</li><li>'What type of fabric is ideal for making casual t-shirts with a vibrant striped pattern?'</li><li>'What is the recommended textile for making versatile garments that can be layered in cooler climates?'</li></ul> |
| Fabric ID 0734 | <ul><li>'What knit fabric is versatile for use in various seasons and holds its shape well?'</li><li>'What type of fabric is recommended for creating casual tops with a gentle, soft drape?'</li><li>'What fabric is suitable for making lightweight and comfortable casual tops for everyday wear?'</li></ul> |
| Fabric ID 0615 | <ul><li>'What fabric would be recommended for making moisture-wicking blouses suitable for warm climates?'</li><li>'What fabric would be apt for creating garments that require a fine, even weave structure?'</li><li>'What is a suitable fabric for creating drapery in light jackets with a slight sheen?'</li></ul> |
| Fabric ID 0869 | <ul><li>'What type of cotton fabric is ideal for making shirts and blouses with a soft drape?'</li><li>'What textile has a slightly textured surface with a fine yet distinct weave?'</li><li>'Which cotton fabric is versatile and suitable for both menswear and womenswear?'</li></ul> |
| Fabric ID 0864 | <ul><li>'Which fabric is breathable and soft to the touch, suitable for creating comfortable dresses?'</li><li>'Which fabric is recommended for making year-round garments with high color saturation?'</li><li>'What fabric can be used for making shirts, pants, and dresses that require a smooth drape and a hint of elasticity?'</li></ul> |
| Fabric ID 0616 | <ul><li>'Which fabric is ideal for creating spring and summer collections with a soft touch and lightweight feel?'</li><li>'What textile is known for its easy care and durability in garment construction?'</li><li>'What type of fabric is best suited for creating blouses with a flowing drape and smooth texture?'</li></ul> |
| Fabric ID 0866 | <ul><li>'Which fabric is durable, resilient, and has a slight give due to the Spandex content?'</li><li>'What fabric has a consistent charcoal gray hue with a matte finish and a twill weave pattern?'</li><li>'What fabric is recommended for making form-fitting jackets that are both durable and breathable?'</li></ul> |
| Fabric ID 0601 | <ul><li>'What fabric would be suitable for creating draped skirts with a smooth surface and stretchability?'</li><li>'What is the best textile for creating draped skirts with a subtle iridescence?'</li><li>'Searching for a fabric with a smooth texture and slight shimmer effect for draped skirts?'</li></ul> |
| Fabric ID 0618 | <ul><li>'What fabric has a soft drape and gentle folds, making it perfect for creating flowy and comfortable spring and summer dresses?'</li><li>'What type of knit fabric offers good resistance to wrinkles and shrinkage for practical everyday wear?'</li><li>'Searching for a polyester knit fabric with a consistent hue and saturation for making versatile and adaptable garments.'</li></ul> |
| Fabric ID 0773 | <ul><li>'Which fabric is versatile and suitable for creating durable garments for everyday wear?'</li><li>'What fabric is suitable for making casual wear like t-shirts, dresses, and tops?'</li><li>'What fabric is known for its stable weave with a small percentage of elastane for comfort and durability?'</li></ul> |
## Evaluation
### Metrics
| Label | Accuracy |
|:--------|:---------|
| **all** | 0.3837 |
## Uses
### Direct Use for Inference
First install the SetFit library:
```bash
pip install setfit
```
Then you can load this model and run inference.
```python
from setfit import SetFitModel
# Download from the 🤗 Hub
model = SetFitModel.from_pretrained("Jazielinho/fabric_model_1")
# Run inference
preds = model("What fabric has a comfortable feel and is suitable for people with sensitive skin?")
```
<!--
### Downstream Use
*List how someone could finetune this model on their own dataset.*
-->
<!--
### Out-of-Scope Use
*List how the model may foreseeably be misused and address what users ought not to do with the model.*
-->
<!--
## Bias, Risks and Limitations
*What are the known or foreseeable issues stemming from this model? You could also flag here known failure cases or weaknesses of the model.*
-->
<!--
### Recommendations
*What are recommendations with respect to the foreseeable issues? For example, filtering explicit content.*
-->
## Training Details
### Training Set Metrics
| Training set | Min | Median | Max |
|:-------------|:----|:--------|:----|
| Word count | 7 | 15.4858 | 30 |
| Label | Training Sample Count |
|:-----------------|:----------------------|
| Fabric ID 0447 | 39 |
| Fabric ID 0448 | 40 |
| Fabric ID 0449 | 41 |
| Fabric ID 0450 | 32 |
| Fabric ID 0453 | 37 |
| Fabric ID 0455 | 33 |
| Fabric ID 0456 | 36 |
| Fabric ID 0458 | 40 |
| Fabric ID 0459 | 30 |
| Fabric ID 0462 | 36 |
| Fabric ID 0462_1 | 42 |
| Fabric ID 0512 | 38 |
| Fabric ID 0513 | 39 |
| Fabric ID 0522 | 43 |
| Fabric ID 0523_1 | 41 |
| Fabric ID 0526_1 | 41 |
| Fabric ID 0527_1 | 35 |
| Fabric ID 0528_1 | 42 |
| Fabric ID 0564 | 40 |
| Fabric ID 0565 | 43 |
| Fabric ID 0571 | 44 |
| Fabric ID 0573 | 36 |
| Fabric ID 0573_1 | 37 |
| Fabric ID 0575 | 40 |
| Fabric ID 0575_1 | 44 |
| Fabric ID 0576_1 | 42 |
| Fabric ID 0578 | 41 |
| Fabric ID 0578_1 | 38 |
| Fabric ID 0579 | 41 |
| Fabric ID 0579_1 | 46 |
| Fabric ID 0596 | 41 |
| Fabric ID 0598 | 38 |
| Fabric ID 0600 | 40 |
| Fabric ID 0601 | 39 |
| Fabric ID 0606 | 41 |
| Fabric ID 0608 | 44 |
| Fabric ID 0612 | 45 |
| Fabric ID 0613 | 40 |
| Fabric ID 0614 | 37 |
| Fabric ID 0615 | 44 |
| Fabric ID 0616 | 39 |
| Fabric ID 0618 | 42 |
| Fabric ID 0645 | 36 |
| Fabric ID 0719_1 | 43 |
| Fabric ID 0722 | 42 |
| Fabric ID 0723 | 37 |
| Fabric ID 0724 | 41 |
| Fabric ID 0725 | 44 |
| Fabric ID 0730 | 36 |
| Fabric ID 0731 | 40 |
| Fabric ID 0733 | 43 |
| Fabric ID 0734 | 44 |
| Fabric ID 0735 | 39 |
| Fabric ID 0736 | 38 |
| Fabric ID 0742 | 38 |
| Fabric ID 0745 | 43 |
| Fabric ID 0748 | 41 |
| Fabric ID 0756 | 44 |
| Fabric ID 0768 | 40 |
| Fabric ID 0769 | 41 |
| Fabric ID 0770 | 35 |
| Fabric ID 0772 | 43 |
| Fabric ID 0773 | 41 |
| Fabric ID 0855 | 43 |
| Fabric ID 0856 | 37 |
| Fabric ID 0859 | 41 |
| Fabric ID 0862 | 36 |
| Fabric ID 0863 | 38 |
| Fabric ID 0864 | 42 |
| Fabric ID 0866 | 41 |
| Fabric ID 0869 | 39 |
| Fabric ID 0873 | 43 |
| Fabric ID 0874 | 34 |
| Fabric ID 0876 | 40 |
| Fabric ID 0880 | 41 |
### Training Hyperparameters
- batch_size: (256, 256)
- num_epochs: (20, 20)
- max_steps: -1
- sampling_strategy: undersampling
- body_learning_rate: (2e-05, 1e-05)
- head_learning_rate: 0.01
- loss: CosineSimilarityLoss
- distance_metric: cosine_distance
- margin: 0.25
- end_to_end: False
- use_amp: False
- warmup_proportion: 0.1
- seed: 42
- eval_max_steps: -1
- load_best_model_at_end: True
### Training Results
| Epoch | Step | Training Loss | Validation Loss |
|:-------:|:--------:|:-------------:|:---------------:|
| 0.0021 | 1 | 0.2732 | - |
| 0.1040 | 50 | 0.2348 | - |
| 0.2079 | 100 | 0.2277 | - |
| 0.3119 | 150 | 0.2419 | - |
| 0.4158 | 200 | 0.2401 | - |
| 0.5198 | 250 | 0.2367 | - |
| 0.6237 | 300 | 0.237 | - |
| 0.7277 | 350 | 0.2372 | - |
| 0.8316 | 400 | 0.2283 | - |
| 0.9356 | 450 | 0.223 | - |
| 1.0 | 481 | - | 0.207 |
| 1.0395 | 500 | 0.2075 | - |
| 1.1435 | 550 | 0.2162 | - |
| 1.2474 | 600 | 0.1984 | - |
| 1.3514 | 650 | 0.2173 | - |
| 1.4553 | 700 | 0.2154 | - |
| 1.5593 | 750 | 0.1912 | - |
| 1.6632 | 800 | 0.2014 | - |
| 1.7672 | 850 | 0.1866 | - |
| 1.8711 | 900 | 0.1933 | - |
| 1.9751 | 950 | 0.1821 | - |
| 2.0 | 962 | - | 0.1863 |
| 2.0790 | 1000 | 0.1607 | - |
| 2.1830 | 1050 | 0.1544 | - |
| 2.2869 | 1100 | 0.1624 | - |
| 2.3909 | 1150 | 0.1586 | - |
| 2.4948 | 1200 | 0.1445 | - |
| 2.5988 | 1250 | 0.1662 | - |
| 2.7027 | 1300 | 0.1515 | - |
| 2.8067 | 1350 | 0.158 | - |
| 2.9106 | 1400 | 0.1316 | - |
| **3.0** | **1443** | **-** | **0.1824** |
| 3.0146 | 1450 | 0.138 | - |
| 3.1185 | 1500 | 0.1414 | - |
| 3.2225 | 1550 | 0.1249 | - |
| 3.3264 | 1600 | 0.1336 | - |
| 3.4304 | 1650 | 0.1249 | - |
| 3.5343 | 1700 | 0.1308 | - |
| 3.6383 | 1750 | 0.1088 | - |
| 3.7422 | 1800 | 0.122 | - |
| 3.8462 | 1850 | 0.1029 | - |
| 3.9501 | 1900 | 0.1065 | - |
| 4.0 | 1924 | - | 0.1836 |
| 4.0541 | 1950 | 0.1133 | - |
| 4.1580 | 2000 | 0.1102 | - |
| 4.2620 | 2050 | 0.1209 | - |
| 4.3659 | 2100 | 0.1054 | - |
| 4.4699 | 2150 | 0.0874 | - |
| 4.5738 | 2200 | 0.0896 | - |
| 4.6778 | 2250 | 0.1104 | - |
| 4.7817 | 2300 | 0.0912 | - |
| 4.8857 | 2350 | 0.0766 | - |
| 4.9896 | 2400 | 0.0778 | - |
| 5.0 | 2405 | - | 0.1952 |
| 5.0936 | 2450 | 0.114 | - |
| 5.1975 | 2500 | 0.0869 | - |
| 5.3015 | 2550 | 0.0912 | - |
| 5.4054 | 2600 | 0.103 | - |
| 5.5094 | 2650 | 0.0748 | - |
| 5.6133 | 2700 | 0.0911 | - |
| 5.7173 | 2750 | 0.0721 | - |
| 5.8212 | 2800 | 0.0964 | - |
| 5.9252 | 2850 | 0.0712 | - |
| 6.0 | 2886 | - | 0.1938 |
| 6.0291 | 2900 | 0.0831 | - |
| 6.1331 | 2950 | 0.0924 | - |
| 6.2370 | 3000 | 0.0862 | - |
| 6.3410 | 3050 | 0.0725 | - |
| 6.4449 | 3100 | 0.0828 | - |
| 6.5489 | 3150 | 0.0645 | - |
| 6.6528 | 3200 | 0.0741 | - |
| 6.7568 | 3250 | 0.0589 | - |
| 6.8607 | 3300 | 0.075 | - |
| 6.9647 | 3350 | 0.075 | - |
| 7.0 | 3367 | - | 0.2016 |
| 7.0686 | 3400 | 0.0893 | - |
| 7.1726 | 3450 | 0.0727 | - |
| 7.2765 | 3500 | 0.0669 | - |
| 7.3805 | 3550 | 0.0702 | - |
| 7.4844 | 3600 | 0.0636 | - |
| 7.5884 | 3650 | 0.0605 | - |
| 7.6923 | 3700 | 0.0707 | - |
| 7.7963 | 3750 | 0.0597 | - |
| 7.9002 | 3800 | 0.0577 | - |
| 8.0 | 3848 | - | 0.2067 |
| 8.0042 | 3850 | 0.0515 | - |
| 8.1081 | 3900 | 0.0686 | - |
| 8.2121 | 3950 | 0.0587 | - |
| 8.3160 | 4000 | 0.057 | - |
| 8.4200 | 4050 | 0.0693 | - |
| 8.5239 | 4100 | 0.0812 | - |
| 8.6279 | 4150 | 0.0592 | - |
| 8.7318 | 4200 | 0.07 | - |
| 8.8358 | 4250 | 0.064 | - |
| 8.9397 | 4300 | 0.0503 | - |
| 9.0 | 4329 | - | 0.2122 |
| 9.0437 | 4350 | 0.0489 | - |
| 9.1476 | 4400 | 0.0602 | - |
| 9.2516 | 4450 | 0.0673 | - |
| 9.3555 | 4500 | 0.0665 | - |
| 9.4595 | 4550 | 0.0672 | - |
| 9.5634 | 4600 | 0.07 | - |
| 9.6674 | 4650 | 0.042 | - |
| 9.7713 | 4700 | 0.0656 | - |
| 9.8753 | 4750 | 0.0557 | - |
| 9.9792 | 4800 | 0.0648 | - |
| 10.0 | 4810 | - | 0.215 |
| 10.0832 | 4850 | 0.0455 | - |
| 10.1871 | 4900 | 0.0668 | - |
| 10.2911 | 4950 | 0.0453 | - |
| 10.3950 | 5000 | 0.0555 | - |
| 10.4990 | 5050 | 0.0679 | - |
| 10.6029 | 5100 | 0.0516 | - |
| 10.7069 | 5150 | 0.0448 | - |
| 10.8108 | 5200 | 0.0458 | - |
| 10.9148 | 5250 | 0.0544 | - |
| 11.0 | 5291 | - | 0.2172 |
| 11.0187 | 5300 | 0.0453 | - |
| 11.1227 | 5350 | 0.0442 | - |
| 11.2266 | 5400 | 0.0396 | - |
| 11.3306 | 5450 | 0.0507 | - |
| 11.4345 | 5500 | 0.0515 | - |
| 11.5385 | 5550 | 0.0503 | - |
| 11.6424 | 5600 | 0.0521 | - |
| 11.7464 | 5650 | 0.0551 | - |
| 11.8503 | 5700 | 0.0572 | - |
| 11.9543 | 5750 | 0.0604 | - |
| 12.0 | 5772 | - | 0.2245 |
| 12.0582 | 5800 | 0.0445 | - |
| 12.1622 | 5850 | 0.0564 | - |
| 12.2661 | 5900 | 0.0449 | - |
| 12.3701 | 5950 | 0.0502 | - |
| 12.4740 | 6000 | 0.0517 | - |
| 12.5780 | 6050 | 0.0426 | - |
| 12.6819 | 6100 | 0.0386 | - |
| 12.7859 | 6150 | 0.0446 | - |
| 12.8898 | 6200 | 0.0574 | - |
| 12.9938 | 6250 | 0.0546 | - |
| 13.0 | 6253 | - | 0.223 |
| 13.0977 | 6300 | 0.0381 | - |
| 13.2017 | 6350 | 0.047 | - |
| 13.3056 | 6400 | 0.0425 | - |
| 13.4096 | 6450 | 0.0445 | - |
| 13.5135 | 6500 | 0.056 | - |
| 13.6175 | 6550 | 0.0533 | - |
| 13.7214 | 6600 | 0.0466 | - |
| 13.8254 | 6650 | 0.0506 | - |
| 13.9293 | 6700 | 0.0402 | - |
| 14.0 | 6734 | - | 0.2238 |
| 14.0333 | 6750 | 0.0375 | - |
| 14.1372 | 6800 | 0.0447 | - |
| 14.2412 | 6850 | 0.0584 | - |
| 14.3451 | 6900 | 0.0348 | - |
| 14.4491 | 6950 | 0.0459 | - |
| 14.5530 | 7000 | 0.0465 | - |
| 14.6570 | 7050 | 0.0421 | - |
| 14.7609 | 7100 | 0.0537 | - |
| 14.8649 | 7150 | 0.041 | - |
| 14.9688 | 7200 | 0.0281 | - |
| 15.0 | 7215 | - | 0.2247 |
| 15.0728 | 7250 | 0.0431 | - |
| 15.1767 | 7300 | 0.039 | - |
| 15.2807 | 7350 | 0.0408 | - |
| 15.3846 | 7400 | 0.048 | - |
| 15.4886 | 7450 | 0.0354 | - |
| 15.5925 | 7500 | 0.0626 | - |
| 15.6965 | 7550 | 0.0396 | - |
| 15.8004 | 7600 | 0.045 | - |
| 15.9044 | 7650 | 0.0432 | - |
| 16.0 | 7696 | - | 0.2246 |
| 16.0083 | 7700 | 0.0385 | - |
| 16.1123 | 7750 | 0.0368 | - |
| 16.2162 | 7800 | 0.0628 | - |
| 16.3202 | 7850 | 0.035 | - |
| 16.4241 | 7900 | 0.0264 | - |
| 16.5281 | 7950 | 0.0275 | - |
| 16.6320 | 8000 | 0.0383 | - |
| 16.7360 | 8050 | 0.0469 | - |
| 16.8399 | 8100 | 0.0445 | - |
| 16.9439 | 8150 | 0.0357 | - |
| 17.0 | 8177 | - | 0.2268 |
| 17.0478 | 8200 | 0.0456 | - |
| 17.1518 | 8250 | 0.053 | - |
| 17.2557 | 8300 | 0.0498 | - |
| 17.3597 | 8350 | 0.0368 | - |
| 17.4636 | 8400 | 0.0473 | - |
| 17.5676 | 8450 | 0.0422 | - |
| 17.6715 | 8500 | 0.0362 | - |
| 17.7755 | 8550 | 0.0292 | - |
| 17.8794 | 8600 | 0.0431 | - |
| 17.9834 | 8650 | 0.0412 | - |
| 18.0 | 8658 | - | 0.2276 |
| 18.0873 | 8700 | 0.0655 | - |
| 18.1913 | 8750 | 0.0405 | - |
| 18.2952 | 8800 | 0.0455 | - |
| 18.3992 | 8850 | 0.0324 | - |
| 18.5031 | 8900 | 0.038 | - |
| 18.6071 | 8950 | 0.0315 | - |
| 18.7110 | 9000 | 0.0468 | - |
| 18.8150 | 9050 | 0.0451 | - |
| 18.9189 | 9100 | 0.032 | - |
| 19.0 | 9139 | - | 0.2268 |
| 19.0229 | 9150 | 0.0371 | - |
| 19.1268 | 9200 | 0.0439 | - |
| 19.2308 | 9250 | 0.0472 | - |
| 19.3347 | 9300 | 0.0362 | - |
| 19.4387 | 9350 | 0.0341 | - |
| 19.5426 | 9400 | 0.036 | - |
| 19.6466 | 9450 | 0.0382 | - |
| 19.7505 | 9500 | 0.0288 | - |
| 19.8545 | 9550 | 0.04 | - |
| 19.9584 | 9600 | 0.0277 | - |
| 20.0 | 9620 | - | 0.2277 |
* The bold row denotes the saved checkpoint.
### Framework Versions
- Python: 3.10.12
- SetFit: 1.0.3
- Sentence Transformers: 2.7.0
- Transformers: 4.40.1
- PyTorch: 2.2.1+cu121
- Datasets: 2.19.0
- Tokenizers: 0.19.1
## Citation
### BibTeX
```bibtex
@article{https://doi.org/10.48550/arxiv.2209.11055,
doi = {10.48550/ARXIV.2209.11055},
url = {https://arxiv.org/abs/2209.11055},
author = {Tunstall, Lewis and Reimers, Nils and Jo, Unso Eun Seo and Bates, Luke and Korat, Daniel and Wasserblat, Moshe and Pereg, Oren},
keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
title = {Efficient Few-Shot Learning Without Prompts},
publisher = {arXiv},
year = {2022},
copyright = {Creative Commons Attribution 4.0 International}
}
```
<!--
## Glossary
*Clearly define terms in order to be accessible across audiences.*
-->
<!--
## Model Card Authors
*Lists the people who create the model card, providing recognition and accountability for the detailed work that goes into its construction.*
-->
<!--
## Model Card Contact
*Provides a way for people who have updates to the Model Card, suggestions, or questions, to contact the Model Card authors.*
--> | {"library_name": "setfit", "tags": ["setfit", "sentence-transformers", "text-classification", "generated_from_setfit_trainer"], "metrics": ["accuracy"], "base_model": "sentence-transformers/paraphrase-MiniLM-L6-v2", "widget": [{"text": "What fabric has a comfortable feel and is suitable for people with sensitive skin?"}, {"text": "What is the most recommended fabric for making outerwear that requires a blend of comfort and resilience?"}, {"text": "What fabric has a fluid drape and is ideal for creating lightweight summer dresses?"}, {"text": "Which fabric is best for creating versatile clothing items like casual shirts, blouses, and dresses in a periwinkle blue hue?"}, {"text": "What kind of fabric is suitable for making form-fitting activewear like yoga pants and t-shirts?"}], "pipeline_tag": "text-classification", "inference": true, "model-index": [{"name": "SetFit with sentence-transformers/paraphrase-MiniLM-L6-v2", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "Unknown", "type": "unknown", "split": "test"}, "metrics": [{"type": "accuracy", "value": 0.3836898395721925, "name": "Accuracy"}]}]}]} | Jazielinho/fabric_model_1 | null | [
"setfit",
"safetensors",
"bert",
"sentence-transformers",
"text-classification",
"generated_from_setfit_trainer",
"arxiv:2209.11055",
"base_model:sentence-transformers/paraphrase-MiniLM-L6-v2",
"model-index",
"region:us"
] | null | 2024-05-01T11:24:33+00:00 |
null | null | {"license": "creativeml-openrail-m"} | casque/011_Pink_gelcoat_v2 | null | [
"license:creativeml-openrail-m",
"region:us"
] | null | 2024-05-01T11:25:33+00:00 |
|
text-generation | transformers |
# Uploaded model
- **Developed by:** herisan
- **License:** apache-2.0
- **Finetuned from model :** unsloth/llama-3-8b-bnb-4bit
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
| {"language": ["en"], "license": "apache-2.0", "tags": ["text-generation-inference", "transformers", "unsloth", "llama", "trl", "sft"], "base_model": "unsloth/llama-3-8b-bnb-4bit"} | herisan/llama-3-8b-alpaca-cleaned | null | [
"transformers",
"safetensors",
"llama",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"sft",
"en",
"base_model:unsloth/llama-3-8b-bnb-4bit",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T11:26:41+00:00 |
token-classification | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# xlm-roberta-base-finetuned-panx-en
This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on an unknown dataset.
It achieves the following results on the evaluation set:
- Loss: 0.3995
- F1: 0.6887
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 5e-05
- train_batch_size: 24
- eval_batch_size: 24
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 3
### Training results
| Training Loss | Epoch | Step | Validation Loss | F1 |
|:-------------:|:-----:|:----:|:---------------:|:------:|
| 1.023 | 1.0 | 50 | 0.5000 | 0.5805 |
| 0.4736 | 2.0 | 100 | 0.4185 | 0.6689 |
| 0.3709 | 3.0 | 150 | 0.3995 | 0.6887 |
### Framework versions
- Transformers 4.39.3
- Pytorch 2.2.2+cu121
- Datasets 2.18.0
- Tokenizers 0.15.2
| {"tags": ["generated_from_trainer"], "metrics": ["f1"], "base_model": "xlm-roberta-base", "model-index": [{"name": "xlm-roberta-base-finetuned-panx-en", "results": []}]} | u00890358/xlm-roberta-base-finetuned-panx-en | null | [
"transformers",
"safetensors",
"xlm-roberta",
"token-classification",
"generated_from_trainer",
"base_model:xlm-roberta-base",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T11:26:50+00:00 |
null | null | {} | umin21/model_out | null | [
"region:us"
] | null | 2024-05-01T11:27:23+00:00 |
|
text-generation | transformers |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# HPY_gpt2_v5
This model is a fine-tuned version of [ClassCat/gpt2-base-french](https://huggingface.co/ClassCat/gpt2-base-french) on the None dataset.
It achieves the following results on the evaluation set:
- Loss: 1.5584
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 2e-05
- train_batch_size: 8
- eval_batch_size: 8
- seed: 42
- gradient_accumulation_steps: 8
- total_train_batch_size: 64
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- num_epochs: 4
### Training results
| Training Loss | Epoch | Step | Validation Loss |
|:-------------:|:-----:|:----:|:---------------:|
| No log | 1.0 | 202 | 1.7781 |
| No log | 2.0 | 404 | 1.6314 |
| 1.9852 | 3.0 | 606 | 1.5753 |
| 1.9852 | 4.0 | 808 | 1.5584 |
### Framework versions
- Transformers 4.30.0
- Pytorch 2.2.1+cu121
- Datasets 2.19.0
- Tokenizers 0.13.3
| {"license": "cc-by-sa-4.0", "tags": ["generated_from_trainer"], "model-index": [{"name": "HPY_gpt2_v5", "results": []}]} | azizkt/HPY_gpt2_v5 | null | [
"transformers",
"pytorch",
"tensorboard",
"gpt2",
"text-generation",
"generated_from_trainer",
"license:cc-by-sa-4.0",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T11:28:10+00:00 |
question-answering | transformers |
<!-- This model card has been generated automatically according to the information Keras had access to. You should
probably proofread and complete it, then remove this comment. -->
# LongRiver/distilbert-base-cased-finetuned
This model is a fine-tuned version of [distilbert-base-cased](https://huggingface.co/distilbert-base-cased) on an unknown dataset.
It achieves the following results on the evaluation set:
- Train Loss: 1.7185
- Train End Logits Accuracy: 0.5917
- Train Start Logits Accuracy: 0.5638
- Validation Loss: 2.0391
- Validation End Logits Accuracy: 0.5252
- Validation Start Logits Accuracy: 0.4886
- Epoch: 1
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- optimizer: {'name': 'Adam', 'weight_decay': None, 'clipnorm': None, 'global_clipnorm': None, 'clipvalue': None, 'use_ema': False, 'ema_momentum': 0.99, 'ema_overwrite_frequency': None, 'jit_compile': True, 'is_legacy_optimizer': False, 'learning_rate': {'module': 'keras.optimizers.schedules', 'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 6786, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered_name': None}, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False}
- training_precision: float32
### Training results
| Train Loss | Train End Logits Accuracy | Train Start Logits Accuracy | Validation Loss | Validation End Logits Accuracy | Validation Start Logits Accuracy | Epoch |
|:----------:|:-------------------------:|:---------------------------:|:---------------:|:------------------------------:|:--------------------------------:|:-----:|
| 2.3543 | 0.5058 | 0.4992 | 2.0820 | 0.5253 | 0.4917 | 0 |
| 1.7185 | 0.5917 | 0.5638 | 2.0391 | 0.5252 | 0.4886 | 1 |
### Framework versions
- Transformers 4.40.1
- TensorFlow 2.15.0
- Datasets 2.19.0
- Tokenizers 0.19.1
| {"license": "apache-2.0", "tags": ["generated_from_keras_callback"], "base_model": "distilbert-base-cased", "model-index": [{"name": "LongRiver/distilbert-base-cased-finetuned", "results": []}]} | LongRiver/distilbert-base-cased-finetuned | null | [
"transformers",
"tf",
"tensorboard",
"distilbert",
"question-answering",
"generated_from_keras_callback",
"base_model:distilbert-base-cased",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T11:29:03+00:00 |
text2text-generation | transformers | {} | abdelmomenn/t5-small-finetuned-emailgen | null | [
"transformers",
"tensorboard",
"safetensors",
"t5",
"text2text-generation",
"autotrain_compatible",
"endpoints_compatible",
"text-generation-inference",
"region:us"
] | null | 2024-05-01T11:30:20+00:00 |
|
token-classification | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {"library_name": "transformers", "tags": []} | 007Rahul/ner_model | null | [
"transformers",
"safetensors",
"bert",
"token-classification",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T11:30:44+00:00 |
null | transformers |
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | {"library_name": "transformers", "tags": []} | 007Rahul/tokenizer | null | [
"transformers",
"arxiv:1910.09700",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T11:31:36+00:00 |
null | transformers | {} | Rasi1610/Death_Se44_newmodel_m9 | null | [
"transformers",
"pytorch",
"vision-encoder-decoder",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T11:32:26+00:00 |
|
null | null | {} | prmditya/bingo | null | [
"region:us"
] | null | 2024-05-01T11:32:44+00:00 |
|
null | null | {} | samuelgaddala/sam | null | [
"region:us"
] | null | 2024-05-01T11:33:21+00:00 |
|
text-generation | transformers |
# Uploaded model
- **Developed by:** abdulrehmanibk
- **License:** apache-2.0
- **Finetuned from model :** unsloth/llama-3-8b-bnb-4bit
This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
[<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth) | {"language": ["en"], "license": "apache-2.0", "tags": ["text-generation-inference", "transformers", "unsloth", "llama", "trl"], "metrics": ["accuracy"], "base_model": "unsloth/llama-3-8b-bnb-4bit", "pipeline_tag": "text-generation"} | abdulrehmanibk/mpg2_project | null | [
"transformers",
"safetensors",
"text-generation-inference",
"unsloth",
"llama",
"trl",
"text-generation",
"en",
"base_model:unsloth/llama-3-8b-bnb-4bit",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | 2024-05-01T11:33:41+00:00 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.