| pipeline_tag
				 stringclasses 48
				values | library_name
				 stringclasses 198
				values | text
				 stringlengths 1 900k | metadata
				 stringlengths 2 438k | id
				 stringlengths 5 122 | last_modified
				 null | tags
				 listlengths 1 1.84k | sha
				 null | created_at
				 stringlengths 25 25 | arxiv
				 listlengths 0 201 | languages
				 listlengths 0 1.83k | tags_str
				 stringlengths 17 9.34k | text_str
				 stringlengths 0 389k | text_lists
				 listlengths 0 722 | processed_texts
				 listlengths 1 723 | tokens_length
				 listlengths 1 723 | input_texts
				 listlengths 1 1 | 
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
| 
	text-generation | 
	transformers | 
	Quantization made by Richard Erkhov.
[Github](https://github.com/RichardErkhov)
[Discord](https://discord.gg/pvy7H8DZMG)
[Request more models](https://github.com/RichardErkhov/quant_request)
sqlcoder-7b-2 - bnb 4bits
- Model creator: https://huggingface.co/defog/
- Original model: https://huggingface.co/defog/sqlcoder-7b-2/
Original model description:
---
license: cc-by-sa-4.0
library_name: transformers
pipeline_tag: text-generation
---
# Update notice
The model weights were updated at 7 AM UTC on Feb 7, 2024. The new model weights lead to a much more performant model – particularly for joins.
If you downloaded the model before that, please redownload the weights for best performance.
# Model Card for SQLCoder-7B-2
A capable large language model for natural language to SQL generation.

## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [Defog, Inc](https://defog.ai)
- **Model type:** [Text to SQL]
- **License:** [CC-by-SA-4.0]
- **Finetuned from model:** [CodeLlama-7B]
### Model Sources [optional]
- [**HuggingFace:**](https://huggingface.co/defog/sqlcoder-70b-alpha)
- [**GitHub:**](https://github.com/defog-ai/sqlcoder)
- [**Demo:**](https://defog.ai/sqlcoder-demo/)
## Uses
This model is intended to be used by non-technical users to understand data inside their SQL databases. It is meant as an analytics tool, and not as a database admin tool.
This model has not been trained to reject malicious requests from users with write access to databases, and should only be used by users with read-only access.
## How to Get Started with the Model
Use the code [here](https://github.com/defog-ai/sqlcoder/blob/main/inference.py) to get started with the model.
## Prompt
Please use the following prompt for optimal results. Please remember to use `do_sample=False` and `num_beams=4` for optimal results.
```
### Task
Generate a SQL query to answer [QUESTION]{user_question}[/QUESTION]
### Database Schema
The query will run on a database with the following schema:
{table_metadata_string_DDL_statements}
### Answer
Given the database schema, here is the SQL query that [QUESTION]{user_question}[/QUESTION]
[SQL]
```
## Evaluation
This model was evaluated on [SQL-Eval](https://github.com/defog-ai/sql-eval), a PostgreSQL based evaluation framework developed by Defog for testing and alignment of model capabilities.
You can read more about the methodology behind SQLEval [here](https://defog.ai/blog/open-sourcing-sqleval/).
### Results
We classified each generated question into one of 6 categories. The table displays the percentage of questions answered correctly by each model, broken down by category.
|                | date | group_by | order_by | ratio | join | where |
| -------------- | ---- | -------- | -------- | ----- | ---- | ----- |
| sqlcoder-70b   | 96   | 91.4     | 97.1     | 85.7  | 97.1 | 91.4  |
| sqlcoder-7b-2  | 96   | 91.4     | 94.3     | 91.4  | 94.3 | 77.1  |
| sqlcoder-34b   | 80   | 94.3     | 85.7     | 77.1  | 85.7 | 80    |
| gpt-4          | 72   | 94.3     | 97.1     | 80    | 91.4 | 80    |
| gpt-4-turbo    | 76   | 91.4     | 91.4     | 62.8  | 88.6 | 77.1  |
| natural-sql-7b | 56   | 88.6     | 85.7     | 60    | 88.6 | 80    |
| sqlcoder-7b    | 64   | 82.9     | 74.3     | 54.3  | 74.3 | 74.3  |
| gpt-3.5        | 72   | 77.1     | 82.8     | 34.3  | 65.7 | 71.4  |
| claude-2       | 52   | 71.4     | 74.3     | 57.1  | 65.7 | 62.9  |
## Model Card Contact
Contact us on X at [@defogdata](https://twitter.com/defogdata), or on email at [[email protected]](mailto:[email protected])
 | 
	{} | 
	RichardErkhov/defog_-_sqlcoder-7b-2-4bits | null | 
	[
  "transformers",
  "safetensors",
  "llama",
  "text-generation",
  "autotrain_compatible",
  "endpoints_compatible",
  "text-generation-inference",
  "4-bit",
  "region:us"
] | null | 
	2024-05-03T19:09:04+00:00 | 
	[] | 
	[] | 
	TAGS
#transformers #safetensors #llama #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us 
 | 
	Quantization made by Richard Erkhov.
Github
Discord
Request more models
sqlcoder-7b-2 - bnb 4bits
* Model creator: URL
* Original model: URL
Original model description:
---------------------------
license: cc-by-sa-4.0
library\_name: transformers
pipeline\_tag: text-generation
--------------------------------------------------------------------------------
Update notice
=============
The model weights were updated at 7 AM UTC on Feb 7, 2024. The new model weights lead to a much more performant model – particularly for joins.
If you downloaded the model before that, please redownload the weights for best performance.
Model Card for SQLCoder-7B-2
============================
A capable large language model for natural language to SQL generation.
!image/png
Model Details
-------------
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
* Developed by: Defog, Inc
* Model type: [Text to SQL]
* License: [CC-by-SA-4.0]
* Finetuned from model: [CodeLlama-7B]
### Model Sources [optional]
* HuggingFace:
* GitHub:
* Demo:
Uses
----
This model is intended to be used by non-technical users to understand data inside their SQL databases. It is meant as an analytics tool, and not as a database admin tool.
This model has not been trained to reject malicious requests from users with write access to databases, and should only be used by users with read-only access.
How to Get Started with the Model
---------------------------------
Use the code here to get started with the model.
Prompt
------
Please use the following prompt for optimal results. Please remember to use 'do\_sample=False' and 'num\_beams=4' for optimal results.
Evaluation
----------
This model was evaluated on SQL-Eval, a PostgreSQL based evaluation framework developed by Defog for testing and alignment of model capabilities.
You can read more about the methodology behind SQLEval here.
### Results
We classified each generated question into one of 6 categories. The table displays the percentage of questions answered correctly by each model, broken down by category.
Model Card Contact
------------------
Contact us on X at @defogdata, or on email at founders@URL
 | 
	[
  "### Model Description\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n\n* Developed by: Defog, Inc\n* Model type: [Text to SQL]\n* License: [CC-by-SA-4.0]\n* Finetuned from model: [CodeLlama-7B]",
  "### Model Sources [optional]\n\n\n* HuggingFace:\n* GitHub:\n* Demo:\n\n\nUses\n----\n\n\nThis model is intended to be used by non-technical users to understand data inside their SQL databases. It is meant as an analytics tool, and not as a database admin tool.\n\n\nThis model has not been trained to reject malicious requests from users with write access to databases, and should only be used by users with read-only access.\n\n\nHow to Get Started with the Model\n---------------------------------\n\n\nUse the code here to get started with the model.\n\n\nPrompt\n------\n\n\nPlease use the following prompt for optimal results. Please remember to use 'do\\_sample=False' and 'num\\_beams=4' for optimal results.\n\n\nEvaluation\n----------\n\n\nThis model was evaluated on SQL-Eval, a PostgreSQL based evaluation framework developed by Defog for testing and alignment of model capabilities.\n\n\nYou can read more about the methodology behind SQLEval here.",
  "### Results\n\n\nWe classified each generated question into one of 6 categories. The table displays the percentage of questions answered correctly by each model, broken down by category.\n\n\n\nModel Card Contact\n------------------\n\n\nContact us on X at @defogdata, or on email at founders@URL"
] | 
	[
  "TAGS\n#transformers #safetensors #llama #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n",
  "### Model Description\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n\n* Developed by: Defog, Inc\n* Model type: [Text to SQL]\n* License: [CC-by-SA-4.0]\n* Finetuned from model: [CodeLlama-7B]",
  "### Model Sources [optional]\n\n\n* HuggingFace:\n* GitHub:\n* Demo:\n\n\nUses\n----\n\n\nThis model is intended to be used by non-technical users to understand data inside their SQL databases. It is meant as an analytics tool, and not as a database admin tool.\n\n\nThis model has not been trained to reject malicious requests from users with write access to databases, and should only be used by users with read-only access.\n\n\nHow to Get Started with the Model\n---------------------------------\n\n\nUse the code here to get started with the model.\n\n\nPrompt\n------\n\n\nPlease use the following prompt for optimal results. Please remember to use 'do\\_sample=False' and 'num\\_beams=4' for optimal results.\n\n\nEvaluation\n----------\n\n\nThis model was evaluated on SQL-Eval, a PostgreSQL based evaluation framework developed by Defog for testing and alignment of model capabilities.\n\n\nYou can read more about the methodology behind SQLEval here.",
  "### Results\n\n\nWe classified each generated question into one of 6 categories. The table displays the percentage of questions answered correctly by each model, broken down by category.\n\n\n\nModel Card Contact\n------------------\n\n\nContact us on X at @defogdata, or on email at founders@URL"
] | 
	[
  38,
  76,
  241,
  73
] | 
	[
  "TAGS\n#transformers #safetensors #llama #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n### Model Description\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n\n* Developed by: Defog, Inc\n* Model type: [Text to SQL]\n* License: [CC-by-SA-4.0]\n* Finetuned from model: [CodeLlama-7B]### Model Sources [optional]\n\n\n* HuggingFace:\n* GitHub:\n* Demo:\n\n\nUses\n----\n\n\nThis model is intended to be used by non-technical users to understand data inside their SQL databases. It is meant as an analytics tool, and not as a database admin tool.\n\n\nThis model has not been trained to reject malicious requests from users with write access to databases, and should only be used by users with read-only access.\n\n\nHow to Get Started with the Model\n---------------------------------\n\n\nUse the code here to get started with the model.\n\n\nPrompt\n------\n\n\nPlease use the following prompt for optimal results. Please remember to use 'do\\_sample=False' and 'num\\_beams=4' for optimal results.\n\n\nEvaluation\n----------\n\n\nThis model was evaluated on SQL-Eval, a PostgreSQL based evaluation framework developed by Defog for testing and alignment of model capabilities.\n\n\nYou can read more about the methodology behind SQLEval here.### Results\n\n\nWe classified each generated question into one of 6 categories. The table displays the percentage of questions answered correctly by each model, broken down by category.\n\n\n\nModel Card Contact\n------------------\n\n\nContact us on X at @defogdata, or on email at founders@URL"
] | 
| null | 
	peft | 
	
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# GUE_tf_4-seqsight_4096_512_15M-L32_f
This model is a fine-tuned version of [mahdibaghbanzadeh/seqsight_4096_512_15M](https://huggingface.co/mahdibaghbanzadeh/seqsight_4096_512_15M) on the [mahdibaghbanzadeh/GUE_tf_4](https://huggingface.co/datasets/mahdibaghbanzadeh/GUE_tf_4) dataset.
It achieves the following results on the evaluation set:
- Loss: 0.3704
- F1 Score: 0.8588
- Accuracy: 0.859
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0005
- train_batch_size: 128
- eval_batch_size: 128
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- training_steps: 10000
### Training results
| Training Loss | Epoch | Step  | Validation Loss | F1 Score | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|:--------:|
| 0.5525        | 1.34  | 200   | 0.5126          | 0.7380   | 0.738    |
| 0.4762        | 2.68  | 400   | 0.4937          | 0.7496   | 0.75     |
| 0.4601        | 4.03  | 600   | 0.4891          | 0.7548   | 0.756    |
| 0.4466        | 5.37  | 800   | 0.4828          | 0.7580   | 0.758    |
| 0.4325        | 6.71  | 1000  | 0.4821          | 0.7670   | 0.769    |
| 0.4221        | 8.05  | 1200  | 0.4624          | 0.7769   | 0.777    |
| 0.416         | 9.4   | 1400  | 0.4501          | 0.7809   | 0.781    |
| 0.4062        | 10.74 | 1600  | 0.4531          | 0.7800   | 0.78     |
| 0.3994        | 12.08 | 1800  | 0.4526          | 0.7831   | 0.784    |
| 0.3951        | 13.42 | 2000  | 0.4485          | 0.7939   | 0.794    |
| 0.3826        | 14.77 | 2200  | 0.4444          | 0.7958   | 0.796    |
| 0.3825        | 16.11 | 2400  | 0.4407          | 0.7955   | 0.796    |
| 0.3734        | 17.45 | 2600  | 0.4475          | 0.7848   | 0.785    |
| 0.367         | 18.79 | 2800  | 0.4480          | 0.7940   | 0.794    |
| 0.3628        | 20.13 | 3000  | 0.4385          | 0.8019   | 0.802    |
| 0.3505        | 21.48 | 3200  | 0.4360          | 0.8079   | 0.808    |
| 0.3513        | 22.82 | 3400  | 0.4419          | 0.8037   | 0.804    |
| 0.345         | 24.16 | 3600  | 0.4359          | 0.8080   | 0.808    |
| 0.3405        | 25.5  | 3800  | 0.4313          | 0.8097   | 0.81     |
| 0.3327        | 26.85 | 4000  | 0.4307          | 0.8130   | 0.813    |
| 0.3347        | 28.19 | 4200  | 0.4333          | 0.7970   | 0.797    |
| 0.319         | 29.53 | 4400  | 0.4489          | 0.8188   | 0.819    |
| 0.3213        | 30.87 | 4600  | 0.4355          | 0.8050   | 0.805    |
| 0.3171        | 32.21 | 4800  | 0.4279          | 0.8090   | 0.809    |
| 0.3143        | 33.56 | 5000  | 0.4330          | 0.8120   | 0.812    |
| 0.3113        | 34.9  | 5200  | 0.4400          | 0.8070   | 0.807    |
| 0.3048        | 36.24 | 5400  | 0.4414          | 0.798    | 0.798    |
| 0.2986        | 37.58 | 5600  | 0.4316          | 0.8146   | 0.815    |
| 0.295         | 38.93 | 5800  | 0.4465          | 0.8040   | 0.804    |
| 0.295         | 40.27 | 6000  | 0.4404          | 0.8098   | 0.81     |
| 0.2883        | 41.61 | 6200  | 0.4515          | 0.8090   | 0.809    |
| 0.2897        | 42.95 | 6400  | 0.4408          | 0.8110   | 0.811    |
| 0.2857        | 44.3  | 6600  | 0.4365          | 0.8145   | 0.815    |
| 0.2787        | 45.64 | 6800  | 0.4331          | 0.8120   | 0.812    |
| 0.2862        | 46.98 | 7000  | 0.4335          | 0.8189   | 0.819    |
| 0.2767        | 48.32 | 7200  | 0.4339          | 0.8148   | 0.815    |
| 0.2712        | 49.66 | 7400  | 0.4270          | 0.8129   | 0.813    |
| 0.2712        | 51.01 | 7600  | 0.4322          | 0.8170   | 0.817    |
| 0.2708        | 52.35 | 7800  | 0.4382          | 0.8198   | 0.82     |
| 0.2644        | 53.69 | 8000  | 0.4400          | 0.8160   | 0.816    |
| 0.2678        | 55.03 | 8200  | 0.4366          | 0.8230   | 0.823    |
| 0.2635        | 56.38 | 8400  | 0.4318          | 0.8229   | 0.823    |
| 0.261         | 57.72 | 8600  | 0.4403          | 0.8178   | 0.818    |
| 0.262         | 59.06 | 8800  | 0.4338          | 0.8179   | 0.818    |
| 0.2617        | 60.4  | 9000  | 0.4364          | 0.8220   | 0.822    |
| 0.2545        | 61.74 | 9200  | 0.4385          | 0.8219   | 0.822    |
| 0.2568        | 63.09 | 9400  | 0.4400          | 0.8289   | 0.829    |
| 0.257         | 64.43 | 9600  | 0.4372          | 0.8239   | 0.824    |
| 0.2581        | 65.77 | 9800  | 0.4372          | 0.8249   | 0.825    |
| 0.2546        | 67.11 | 10000 | 0.4370          | 0.8259   | 0.826    |
### Framework versions
- PEFT 0.9.0
- Transformers 4.38.2
- Pytorch 2.2.0+cu121
- Datasets 2.17.1
- Tokenizers 0.15.2 | 
	{"library_name": "peft", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "mahdibaghbanzadeh/seqsight_4096_512_15M", "model-index": [{"name": "GUE_tf_4-seqsight_4096_512_15M-L32_f", "results": []}]} | 
	mahdibaghbanzadeh/GUE_tf_4-seqsight_4096_512_15M-L32_f | null | 
	[
  "peft",
  "safetensors",
  "generated_from_trainer",
  "base_model:mahdibaghbanzadeh/seqsight_4096_512_15M",
  "region:us"
] | null | 
	2024-05-03T19:09:48+00:00 | 
	[] | 
	[] | 
	TAGS
#peft #safetensors #generated_from_trainer #base_model-mahdibaghbanzadeh/seqsight_4096_512_15M #region-us 
 | 
	GUE\_tf\_4-seqsight\_4096\_512\_15M-L32\_f
==========================================
This model is a fine-tuned version of mahdibaghbanzadeh/seqsight\_4096\_512\_15M on the mahdibaghbanzadeh/GUE\_tf\_4 dataset.
It achieves the following results on the evaluation set:
* Loss: 0.3704
* F1 Score: 0.8588
* Accuracy: 0.859
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0005
* train\_batch\_size: 128
* eval\_batch\_size: 128
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* training\_steps: 10000
### Training results
### Framework versions
* PEFT 0.9.0
* Transformers 4.38.2
* Pytorch 2.2.0+cu121
* Datasets 2.17.1
* Tokenizers 0.15.2
 | 
	[
  "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000",
  "### Training results",
  "### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
	[
  "TAGS\n#peft #safetensors #generated_from_trainer #base_model-mahdibaghbanzadeh/seqsight_4096_512_15M #region-us \n",
  "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000",
  "### Training results",
  "### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
	[
  43,
  100,
  5,
  52
] | 
	[
  "TAGS\n#peft #safetensors #generated_from_trainer #base_model-mahdibaghbanzadeh/seqsight_4096_512_15M #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000### Training results### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
| null | 
	peft | 
	
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# GUE_tf_3-seqsight_4096_512_15M-L1_f
This model is a fine-tuned version of [mahdibaghbanzadeh/seqsight_4096_512_15M](https://huggingface.co/mahdibaghbanzadeh/seqsight_4096_512_15M) on the [mahdibaghbanzadeh/GUE_tf_3](https://huggingface.co/datasets/mahdibaghbanzadeh/GUE_tf_3) dataset.
It achieves the following results on the evaluation set:
- Loss: 0.5670
- F1 Score: 0.6927
- Accuracy: 0.696
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0005
- train_batch_size: 128
- eval_batch_size: 128
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- training_steps: 10000
### Training results
| Training Loss | Epoch | Step  | Validation Loss | F1 Score | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|:--------:|
| 0.6616        | 0.93  | 200   | 0.6000          | 0.6809   | 0.682    |
| 0.618         | 1.87  | 400   | 0.5894          | 0.6799   | 0.68     |
| 0.6049        | 2.8   | 600   | 0.5722          | 0.7073   | 0.711    |
| 0.6017        | 3.74  | 800   | 0.5692          | 0.7085   | 0.71     |
| 0.5991        | 4.67  | 1000  | 0.5638          | 0.7147   | 0.716    |
| 0.5926        | 5.61  | 1200  | 0.5632          | 0.7197   | 0.72     |
| 0.5904        | 6.54  | 1400  | 0.5591          | 0.7143   | 0.716    |
| 0.5889        | 7.48  | 1600  | 0.5586          | 0.7229   | 0.723    |
| 0.5877        | 8.41  | 1800  | 0.5571          | 0.7163   | 0.717    |
| 0.59          | 9.35  | 2000  | 0.5569          | 0.7177   | 0.719    |
| 0.5839        | 10.28 | 2200  | 0.5590          | 0.7100   | 0.71     |
| 0.5842        | 11.21 | 2400  | 0.5519          | 0.7183   | 0.72     |
| 0.5841        | 12.15 | 2600  | 0.5506          | 0.7176   | 0.721    |
| 0.581         | 13.08 | 2800  | 0.5494          | 0.7161   | 0.719    |
| 0.5822        | 14.02 | 3000  | 0.5530          | 0.7166   | 0.717    |
| 0.5808        | 14.95 | 3200  | 0.5503          | 0.7212   | 0.722    |
| 0.5786        | 15.89 | 3400  | 0.5493          | 0.7234   | 0.725    |
| 0.5755        | 16.82 | 3600  | 0.5515          | 0.7176   | 0.718    |
| 0.5761        | 17.76 | 3800  | 0.5495          | 0.7271   | 0.729    |
| 0.5766        | 18.69 | 4000  | 0.5525          | 0.7197   | 0.72     |
| 0.5732        | 19.63 | 4200  | 0.5478          | 0.7169   | 0.721    |
| 0.5766        | 20.56 | 4400  | 0.5462          | 0.7184   | 0.72     |
| 0.5746        | 21.5  | 4600  | 0.5500          | 0.7120   | 0.712    |
| 0.5734        | 22.43 | 4800  | 0.5467          | 0.7263   | 0.728    |
| 0.5739        | 23.36 | 5000  | 0.5478          | 0.7246   | 0.725    |
| 0.5734        | 24.3  | 5200  | 0.5494          | 0.7121   | 0.712    |
| 0.5696        | 25.23 | 5400  | 0.5453          | 0.7188   | 0.722    |
| 0.5745        | 26.17 | 5600  | 0.5448          | 0.7234   | 0.725    |
| 0.568         | 27.1  | 5800  | 0.5439          | 0.7209   | 0.724    |
| 0.5682        | 28.04 | 6000  | 0.5437          | 0.7299   | 0.731    |
| 0.569         | 28.97 | 6200  | 0.5486          | 0.7161   | 0.716    |
| 0.5717        | 29.91 | 6400  | 0.5448          | 0.7316   | 0.733    |
| 0.5681        | 30.84 | 6600  | 0.5447          | 0.7337   | 0.735    |
| 0.5686        | 31.78 | 6800  | 0.5464          | 0.7217   | 0.722    |
| 0.5681        | 32.71 | 7000  | 0.5444          | 0.7319   | 0.733    |
| 0.5714        | 33.64 | 7200  | 0.5447          | 0.7315   | 0.733    |
| 0.5642        | 34.58 | 7400  | 0.5480          | 0.7131   | 0.713    |
| 0.5704        | 35.51 | 7600  | 0.5458          | 0.7226   | 0.723    |
| 0.5689        | 36.45 | 7800  | 0.5453          | 0.7246   | 0.725    |
| 0.5676        | 37.38 | 8000  | 0.5453          | 0.7236   | 0.724    |
| 0.5647        | 38.32 | 8200  | 0.5449          | 0.7317   | 0.733    |
| 0.5652        | 39.25 | 8400  | 0.5451          | 0.7284   | 0.729    |
| 0.5662        | 40.19 | 8600  | 0.5453          | 0.7284   | 0.729    |
| 0.5649        | 41.12 | 8800  | 0.5455          | 0.7275   | 0.728    |
| 0.5682        | 42.06 | 9000  | 0.5454          | 0.7285   | 0.729    |
| 0.5665        | 42.99 | 9200  | 0.5461          | 0.7217   | 0.722    |
| 0.565         | 43.93 | 9400  | 0.5464          | 0.7199   | 0.72     |
| 0.5637        | 44.86 | 9600  | 0.5452          | 0.7266   | 0.727    |
| 0.5659        | 45.79 | 9800  | 0.5451          | 0.7285   | 0.729    |
| 0.562         | 46.73 | 10000 | 0.5452          | 0.7256   | 0.726    |
### Framework versions
- PEFT 0.9.0
- Transformers 4.38.2
- Pytorch 2.2.0+cu121
- Datasets 2.17.1
- Tokenizers 0.15.2 | 
	{"library_name": "peft", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "mahdibaghbanzadeh/seqsight_4096_512_15M", "model-index": [{"name": "GUE_tf_3-seqsight_4096_512_15M-L1_f", "results": []}]} | 
	mahdibaghbanzadeh/GUE_tf_3-seqsight_4096_512_15M-L1_f | null | 
	[
  "peft",
  "safetensors",
  "generated_from_trainer",
  "base_model:mahdibaghbanzadeh/seqsight_4096_512_15M",
  "region:us"
] | null | 
	2024-05-03T19:09:48+00:00 | 
	[] | 
	[] | 
	TAGS
#peft #safetensors #generated_from_trainer #base_model-mahdibaghbanzadeh/seqsight_4096_512_15M #region-us 
 | 
	GUE\_tf\_3-seqsight\_4096\_512\_15M-L1\_f
=========================================
This model is a fine-tuned version of mahdibaghbanzadeh/seqsight\_4096\_512\_15M on the mahdibaghbanzadeh/GUE\_tf\_3 dataset.
It achieves the following results on the evaluation set:
* Loss: 0.5670
* F1 Score: 0.6927
* Accuracy: 0.696
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0005
* train\_batch\_size: 128
* eval\_batch\_size: 128
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* training\_steps: 10000
### Training results
### Framework versions
* PEFT 0.9.0
* Transformers 4.38.2
* Pytorch 2.2.0+cu121
* Datasets 2.17.1
* Tokenizers 0.15.2
 | 
	[
  "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000",
  "### Training results",
  "### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
	[
  "TAGS\n#peft #safetensors #generated_from_trainer #base_model-mahdibaghbanzadeh/seqsight_4096_512_15M #region-us \n",
  "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000",
  "### Training results",
  "### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
	[
  43,
  100,
  5,
  52
] | 
	[
  "TAGS\n#peft #safetensors #generated_from_trainer #base_model-mahdibaghbanzadeh/seqsight_4096_512_15M #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000### Training results### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
| null | 
	peft | 
	
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# GUE_tf_3-seqsight_4096_512_15M-L8_f
This model is a fine-tuned version of [mahdibaghbanzadeh/seqsight_4096_512_15M](https://huggingface.co/mahdibaghbanzadeh/seqsight_4096_512_15M) on the [mahdibaghbanzadeh/GUE_tf_3](https://huggingface.co/datasets/mahdibaghbanzadeh/GUE_tf_3) dataset.
It achieves the following results on the evaluation set:
- Loss: 0.5516
- F1 Score: 0.7033
- Accuracy: 0.706
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0005
- train_batch_size: 128
- eval_batch_size: 128
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- training_steps: 10000
### Training results
| Training Loss | Epoch | Step  | Validation Loss | F1 Score | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|:--------:|
| 0.6433        | 0.93  | 200   | 0.5805          | 0.7011   | 0.701    |
| 0.6017        | 1.87  | 400   | 0.5751          | 0.7004   | 0.701    |
| 0.594         | 2.8   | 600   | 0.5608          | 0.7085   | 0.711    |
| 0.5899        | 3.74  | 800   | 0.5582          | 0.7090   | 0.709    |
| 0.5889        | 4.67  | 1000  | 0.5522          | 0.7134   | 0.714    |
| 0.5826        | 5.61  | 1200  | 0.5491          | 0.7141   | 0.715    |
| 0.5801        | 6.54  | 1400  | 0.5494          | 0.7171   | 0.718    |
| 0.5778        | 7.48  | 1600  | 0.5482          | 0.7223   | 0.723    |
| 0.5758        | 8.41  | 1800  | 0.5475          | 0.7218   | 0.722    |
| 0.5787        | 9.35  | 2000  | 0.5472          | 0.7054   | 0.709    |
| 0.5717        | 10.28 | 2200  | 0.5482          | 0.7199   | 0.72     |
| 0.5721        | 11.21 | 2400  | 0.5441          | 0.7227   | 0.724    |
| 0.5709        | 12.15 | 2600  | 0.5453          | 0.7008   | 0.707    |
| 0.5673        | 13.08 | 2800  | 0.5479          | 0.6937   | 0.701    |
| 0.5676        | 14.02 | 3000  | 0.5444          | 0.7196   | 0.721    |
| 0.5661        | 14.95 | 3200  | 0.5459          | 0.7086   | 0.712    |
| 0.5641        | 15.89 | 3400  | 0.5448          | 0.7142   | 0.716    |
| 0.5601        | 16.82 | 3600  | 0.5457          | 0.7172   | 0.719    |
| 0.5597        | 17.76 | 3800  | 0.5455          | 0.7127   | 0.716    |
| 0.5602        | 18.69 | 4000  | 0.5471          | 0.7187   | 0.719    |
| 0.558         | 19.63 | 4200  | 0.5495          | 0.7043   | 0.709    |
| 0.559         | 20.56 | 4400  | 0.5477          | 0.7125   | 0.716    |
| 0.5577        | 21.5  | 4600  | 0.5518          | 0.7161   | 0.716    |
| 0.5555        | 22.43 | 4800  | 0.5469          | 0.7103   | 0.714    |
| 0.5556        | 23.36 | 5000  | 0.5495          | 0.7171   | 0.717    |
| 0.5544        | 24.3  | 5200  | 0.5554          | 0.6955   | 0.696    |
| 0.5502        | 25.23 | 5400  | 0.5482          | 0.7157   | 0.719    |
| 0.5575        | 26.17 | 5600  | 0.5434          | 0.7264   | 0.728    |
| 0.5477        | 27.1  | 5800  | 0.5433          | 0.7174   | 0.719    |
| 0.5481        | 28.04 | 6000  | 0.5441          | 0.7282   | 0.73     |
| 0.5482        | 28.97 | 6200  | 0.5480          | 0.7231   | 0.723    |
| 0.5491        | 29.91 | 6400  | 0.5455          | 0.7245   | 0.727    |
| 0.5473        | 30.84 | 6600  | 0.5441          | 0.7217   | 0.723    |
| 0.5492        | 31.78 | 6800  | 0.5472          | 0.7217   | 0.722    |
| 0.5466        | 32.71 | 7000  | 0.5442          | 0.7272   | 0.728    |
| 0.5503        | 33.64 | 7200  | 0.5444          | 0.7283   | 0.73     |
| 0.542         | 34.58 | 7400  | 0.5502          | 0.7191   | 0.719    |
| 0.5477        | 35.51 | 7600  | 0.5458          | 0.7290   | 0.729    |
| 0.5467        | 36.45 | 7800  | 0.5461          | 0.7257   | 0.726    |
| 0.5466        | 37.38 | 8000  | 0.5456          | 0.7278   | 0.728    |
| 0.5417        | 38.32 | 8200  | 0.5471          | 0.7259   | 0.727    |
| 0.5427        | 39.25 | 8400  | 0.5465          | 0.7237   | 0.724    |
| 0.5423        | 40.19 | 8600  | 0.5461          | 0.7255   | 0.726    |
| 0.5414        | 41.12 | 8800  | 0.5461          | 0.7285   | 0.729    |
| 0.5451        | 42.06 | 9000  | 0.5452          | 0.7277   | 0.728    |
| 0.5428        | 42.99 | 9200  | 0.5468          | 0.7259   | 0.726    |
| 0.541         | 43.93 | 9400  | 0.5469          | 0.7259   | 0.726    |
| 0.538         | 44.86 | 9600  | 0.5463          | 0.7257   | 0.726    |
| 0.5423        | 45.79 | 9800  | 0.5461          | 0.7293   | 0.73     |
| 0.5373        | 46.73 | 10000 | 0.5468          | 0.7248   | 0.725    |
### Framework versions
- PEFT 0.9.0
- Transformers 4.38.2
- Pytorch 2.2.0+cu121
- Datasets 2.17.1
- Tokenizers 0.15.2 | 
	{"library_name": "peft", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "mahdibaghbanzadeh/seqsight_4096_512_15M", "model-index": [{"name": "GUE_tf_3-seqsight_4096_512_15M-L8_f", "results": []}]} | 
	mahdibaghbanzadeh/GUE_tf_3-seqsight_4096_512_15M-L8_f | null | 
	[
  "peft",
  "safetensors",
  "generated_from_trainer",
  "base_model:mahdibaghbanzadeh/seqsight_4096_512_15M",
  "region:us"
] | null | 
	2024-05-03T19:10:15+00:00 | 
	[] | 
	[] | 
	TAGS
#peft #safetensors #generated_from_trainer #base_model-mahdibaghbanzadeh/seqsight_4096_512_15M #region-us 
 | 
	GUE\_tf\_3-seqsight\_4096\_512\_15M-L8\_f
=========================================
This model is a fine-tuned version of mahdibaghbanzadeh/seqsight\_4096\_512\_15M on the mahdibaghbanzadeh/GUE\_tf\_3 dataset.
It achieves the following results on the evaluation set:
* Loss: 0.5516
* F1 Score: 0.7033
* Accuracy: 0.706
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0005
* train\_batch\_size: 128
* eval\_batch\_size: 128
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* training\_steps: 10000
### Training results
### Framework versions
* PEFT 0.9.0
* Transformers 4.38.2
* Pytorch 2.2.0+cu121
* Datasets 2.17.1
* Tokenizers 0.15.2
 | 
	[
  "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000",
  "### Training results",
  "### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
	[
  "TAGS\n#peft #safetensors #generated_from_trainer #base_model-mahdibaghbanzadeh/seqsight_4096_512_15M #region-us \n",
  "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000",
  "### Training results",
  "### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
	[
  43,
  100,
  5,
  52
] | 
	[
  "TAGS\n#peft #safetensors #generated_from_trainer #base_model-mahdibaghbanzadeh/seqsight_4096_512_15M #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000### Training results### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
| null | null | 
	
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# GUE_tf_2-seqsight_4096_512_15M-L1_f
This model is a fine-tuned version of [mahdibaghbanzadeh/seqsight_4096_512_15M](https://huggingface.co/mahdibaghbanzadeh/seqsight_4096_512_15M) on the [mahdibaghbanzadeh/GUE_tf_2](https://huggingface.co/datasets/mahdibaghbanzadeh/GUE_tf_2) dataset.
It achieves the following results on the evaluation set:
- Loss: 0.4762
- F1 Score: 0.7710
- Accuracy: 0.771
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0005
- train_batch_size: 128
- eval_batch_size: 128
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- training_steps: 10000
### Training results
| Training Loss | Epoch | Step  | Validation Loss | F1 Score | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|:--------:|
| 0.6238        | 1.34  | 200   | 0.5599          | 0.7084   | 0.71     |
| 0.5565        | 2.68  | 400   | 0.5340          | 0.7329   | 0.733    |
| 0.538         | 4.03  | 600   | 0.5271          | 0.7310   | 0.731    |
| 0.5338        | 5.37  | 800   | 0.5242          | 0.7350   | 0.735    |
| 0.5298        | 6.71  | 1000  | 0.5203          | 0.7365   | 0.737    |
| 0.5247        | 8.05  | 1200  | 0.5171          | 0.7490   | 0.749    |
| 0.5214        | 9.4   | 1400  | 0.5140          | 0.7415   | 0.742    |
| 0.5209        | 10.74 | 1600  | 0.5141          | 0.7418   | 0.742    |
| 0.5181        | 12.08 | 1800  | 0.5195          | 0.7438   | 0.744    |
| 0.5183        | 13.42 | 2000  | 0.5135          | 0.7440   | 0.744    |
| 0.5182        | 14.77 | 2200  | 0.5122          | 0.7470   | 0.747    |
| 0.5123        | 16.11 | 2400  | 0.5162          | 0.7407   | 0.741    |
| 0.5154        | 17.45 | 2600  | 0.5111          | 0.7399   | 0.74     |
| 0.5098        | 18.79 | 2800  | 0.5099          | 0.7400   | 0.74     |
| 0.5091        | 20.13 | 3000  | 0.5103          | 0.7400   | 0.74     |
| 0.5095        | 21.48 | 3200  | 0.5116          | 0.7359   | 0.736    |
| 0.5106        | 22.82 | 3400  | 0.5074          | 0.7399   | 0.74     |
| 0.5052        | 24.16 | 3600  | 0.5060          | 0.7358   | 0.736    |
| 0.5024        | 25.5  | 3800  | 0.5064          | 0.7342   | 0.735    |
| 0.505         | 26.85 | 4000  | 0.5060          | 0.7375   | 0.738    |
| 0.5014        | 28.19 | 4200  | 0.5058          | 0.7340   | 0.734    |
| 0.5024        | 29.53 | 4400  | 0.5097          | 0.7410   | 0.741    |
| 0.5034        | 30.87 | 4600  | 0.5076          | 0.7380   | 0.738    |
| 0.5015        | 32.21 | 4800  | 0.5058          | 0.7390   | 0.739    |
| 0.5012        | 33.56 | 5000  | 0.5107          | 0.7417   | 0.742    |
| 0.5032        | 34.9  | 5200  | 0.5063          | 0.7389   | 0.739    |
| 0.4975        | 36.24 | 5400  | 0.5017          | 0.7367   | 0.737    |
| 0.4993        | 37.58 | 5600  | 0.5034          | 0.7420   | 0.742    |
| 0.4966        | 38.93 | 5800  | 0.5047          | 0.7370   | 0.737    |
| 0.497         | 40.27 | 6000  | 0.5033          | 0.7360   | 0.736    |
| 0.4973        | 41.61 | 6200  | 0.5028          | 0.7320   | 0.732    |
| 0.4951        | 42.95 | 6400  | 0.5043          | 0.7340   | 0.734    |
| 0.4949        | 44.3  | 6600  | 0.5056          | 0.7370   | 0.737    |
| 0.4977        | 45.64 | 6800  | 0.5057          | 0.7420   | 0.742    |
| 0.4943        | 46.98 | 7000  | 0.5042          | 0.7400   | 0.74     |
| 0.4949        | 48.32 | 7200  | 0.5059          | 0.7380   | 0.738    |
| 0.4923        | 49.66 | 7400  | 0.5017          | 0.7390   | 0.739    |
| 0.4941        | 51.01 | 7600  | 0.5031          | 0.7400   | 0.74     |
| 0.4942        | 52.35 | 7800  | 0.5022          | 0.7390   | 0.739    |
| 0.4957        | 53.69 | 8000  | 0.5019          | 0.7299   | 0.73     |
| 0.492         | 55.03 | 8200  | 0.5023          | 0.7410   | 0.741    |
| 0.4959        | 56.38 | 8400  | 0.5038          | 0.7400   | 0.74     |
| 0.494         | 57.72 | 8600  | 0.5026          | 0.7370   | 0.737    |
| 0.4905        | 59.06 | 8800  | 0.5026          | 0.7340   | 0.734    |
| 0.4909        | 60.4  | 9000  | 0.5039          | 0.7390   | 0.739    |
| 0.4921        | 61.74 | 9200  | 0.5022          | 0.7360   | 0.736    |
| 0.4956        | 63.09 | 9400  | 0.5020          | 0.7360   | 0.736    |
| 0.4896        | 64.43 | 9600  | 0.5025          | 0.7380   | 0.738    |
| 0.4913        | 65.77 | 9800  | 0.5032          | 0.7370   | 0.737    |
| 0.4887        | 67.11 | 10000 | 0.5025          | 0.7370   | 0.737    |
### Framework versions
- PEFT 0.9.0
- Transformers 4.38.2
- Pytorch 2.2.0+cu121
- Datasets 2.17.1
- Tokenizers 0.15.2 | 
	{"library_name": "peft", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "mahdibaghbanzadeh/seqsight_4096_512_15M", "model-index": [{"name": "GUE_tf_2-seqsight_4096_512_15M-L1_f", "results": []}]} | 
	mahdibaghbanzadeh/GUE_tf_2-seqsight_4096_512_15M-L1_f | null | 
	[
  "region:us"
] | null | 
	2024-05-03T19:11:10+00:00 | 
	[] | 
	[] | 
	TAGS
#region-us 
 | 
	GUE\_tf\_2-seqsight\_4096\_512\_15M-L1\_f
=========================================
This model is a fine-tuned version of mahdibaghbanzadeh/seqsight\_4096\_512\_15M on the mahdibaghbanzadeh/GUE\_tf\_2 dataset.
It achieves the following results on the evaluation set:
* Loss: 0.4762
* F1 Score: 0.7710
* Accuracy: 0.771
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0005
* train\_batch\_size: 128
* eval\_batch\_size: 128
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* training\_steps: 10000
### Training results
### Framework versions
* PEFT 0.9.0
* Transformers 4.38.2
* Pytorch 2.2.0+cu121
* Datasets 2.17.1
* Tokenizers 0.15.2
 | 
	[
  "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000",
  "### Training results",
  "### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
	[
  "TAGS\n#region-us \n",
  "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000",
  "### Training results",
  "### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
	[
  5,
  100,
  5,
  52
] | 
	[
  "TAGS\n#region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000### Training results### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
| null | 
	transformers | 
	
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure 
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
 | 
	{"library_name": "transformers", "tags": []} | 
	ferrazzipietro/LS_Llama-2-7b-hf_adapters_en.layer1_NoQuant_32_32_0.05_2_5e-05 | null | 
	[
  "transformers",
  "safetensors",
  "arxiv:1910.09700",
  "endpoints_compatible",
  "region:us"
] | null | 
	2024-05-03T19:11:20+00:00 | 
	[
  "1910.09700"
] | 
	[] | 
	TAGS
#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us 
 | 
	
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a  transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by: 
- Funded by [optional]: 
- Shared by [optional]: 
- Model type: 
- Language(s) (NLP): 
- License: 
- Finetuned from model [optional]: 
### Model Sources [optional]
- Repository: 
- Paper [optional]: 
- Demo [optional]: 
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure 
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:  
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type: 
- Hours used: 
- Cloud Provider: 
- Compute Region: 
- Carbon Emitted: 
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
 | 
	[
  "# Model Card for Model ID",
  "## Model Details",
  "### Model Description\n\n\n\nThis is the model card of a  transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
  "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
  "## Uses",
  "### Direct Use",
  "### Downstream Use [optional]",
  "### Out-of-Scope Use",
  "## Bias, Risks, and Limitations",
  "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
  "## How to Get Started with the Model\n\nUse the code below to get started with the model.",
  "## Training Details",
  "### Training Data",
  "### Training Procedure",
  "#### Preprocessing [optional]",
  "#### Training Hyperparameters\n\n- Training regime:",
  "#### Speeds, Sizes, Times [optional]",
  "## Evaluation",
  "### Testing Data, Factors & Metrics",
  "#### Testing Data",
  "#### Factors",
  "#### Metrics",
  "### Results",
  "#### Summary",
  "## Model Examination [optional]",
  "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
  "## Technical Specifications [optional]",
  "### Model Architecture and Objective",
  "### Compute Infrastructure",
  "#### Hardware",
  "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
  "## Glossary [optional]",
  "## More Information [optional]",
  "## Model Card Authors [optional]",
  "## Model Card Contact"
] | 
	[
  "TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n",
  "# Model Card for Model ID",
  "## Model Details",
  "### Model Description\n\n\n\nThis is the model card of a  transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
  "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
  "## Uses",
  "### Direct Use",
  "### Downstream Use [optional]",
  "### Out-of-Scope Use",
  "## Bias, Risks, and Limitations",
  "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
  "## How to Get Started with the Model\n\nUse the code below to get started with the model.",
  "## Training Details",
  "### Training Data",
  "### Training Procedure",
  "#### Preprocessing [optional]",
  "#### Training Hyperparameters\n\n- Training regime:",
  "#### Speeds, Sizes, Times [optional]",
  "## Evaluation",
  "### Testing Data, Factors & Metrics",
  "#### Testing Data",
  "#### Factors",
  "#### Metrics",
  "### Results",
  "#### Summary",
  "## Model Examination [optional]",
  "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
  "## Technical Specifications [optional]",
  "### Model Architecture and Objective",
  "### Compute Infrastructure",
  "#### Hardware",
  "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
  "## Glossary [optional]",
  "## More Information [optional]",
  "## Model Card Authors [optional]",
  "## Model Card Contact"
] | 
	[
  26,
  6,
  4,
  75,
  23,
  3,
  5,
  8,
  9,
  8,
  34,
  20,
  4,
  5,
  5,
  11,
  13,
  12,
  3,
  10,
  6,
  5,
  6,
  4,
  5,
  7,
  49,
  7,
  7,
  5,
  5,
  15,
  7,
  7,
  8,
  5
] | 
	[
  "TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a  transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | 
| null | 
	transformers | 
	
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed] | 
	{"library_name": "transformers", "tags": []} | 
	Armandodelca/Prototipo_7_EMI | null | 
	[
  "transformers",
  "arxiv:1910.09700",
  "endpoints_compatible",
  "region:us"
] | null | 
	2024-05-03T19:12:51+00:00 | 
	[
  "1910.09700"
] | 
	[] | 
	TAGS
#transformers #arxiv-1910.09700 #endpoints_compatible #region-us 
 | 
	
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a  transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by: 
- Funded by [optional]: 
- Shared by [optional]: 
- Model type: 
- Language(s) (NLP): 
- License: 
- Finetuned from model [optional]: 
### Model Sources [optional]
- Repository: 
- Paper [optional]: 
- Demo [optional]: 
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:  
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type: 
- Hours used: 
- Cloud Provider: 
- Compute Region: 
- Carbon Emitted: 
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
 | 
	[
  "# Model Card for Model ID",
  "## Model Details",
  "### Model Description\n\n\n\nThis is the model card of a  transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
  "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
  "## Uses",
  "### Direct Use",
  "### Downstream Use [optional]",
  "### Out-of-Scope Use",
  "## Bias, Risks, and Limitations",
  "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
  "## How to Get Started with the Model\n\nUse the code below to get started with the model.",
  "## Training Details",
  "### Training Data",
  "### Training Procedure",
  "#### Preprocessing [optional]",
  "#### Training Hyperparameters\n\n- Training regime:",
  "#### Speeds, Sizes, Times [optional]",
  "## Evaluation",
  "### Testing Data, Factors & Metrics",
  "#### Testing Data",
  "#### Factors",
  "#### Metrics",
  "### Results",
  "#### Summary",
  "## Model Examination [optional]",
  "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
  "## Technical Specifications [optional]",
  "### Model Architecture and Objective",
  "### Compute Infrastructure",
  "#### Hardware",
  "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
  "## Glossary [optional]",
  "## More Information [optional]",
  "## Model Card Authors [optional]",
  "## Model Card Contact"
] | 
	[
  "TAGS\n#transformers #arxiv-1910.09700 #endpoints_compatible #region-us \n",
  "# Model Card for Model ID",
  "## Model Details",
  "### Model Description\n\n\n\nThis is the model card of a  transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
  "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
  "## Uses",
  "### Direct Use",
  "### Downstream Use [optional]",
  "### Out-of-Scope Use",
  "## Bias, Risks, and Limitations",
  "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
  "## How to Get Started with the Model\n\nUse the code below to get started with the model.",
  "## Training Details",
  "### Training Data",
  "### Training Procedure",
  "#### Preprocessing [optional]",
  "#### Training Hyperparameters\n\n- Training regime:",
  "#### Speeds, Sizes, Times [optional]",
  "## Evaluation",
  "### Testing Data, Factors & Metrics",
  "#### Testing Data",
  "#### Factors",
  "#### Metrics",
  "### Results",
  "#### Summary",
  "## Model Examination [optional]",
  "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
  "## Technical Specifications [optional]",
  "### Model Architecture and Objective",
  "### Compute Infrastructure",
  "#### Hardware",
  "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
  "## Glossary [optional]",
  "## More Information [optional]",
  "## Model Card Authors [optional]",
  "## Model Card Contact"
] | 
	[
  22,
  6,
  4,
  75,
  23,
  3,
  5,
  8,
  9,
  8,
  34,
  20,
  4,
  5,
  5,
  11,
  13,
  12,
  3,
  10,
  6,
  5,
  6,
  4,
  5,
  7,
  49,
  7,
  7,
  5,
  5,
  15,
  7,
  7,
  8,
  5
] | 
	[
  "TAGS\n#transformers #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a  transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | 
| null | null | 
	Quantization made by Richard Erkhov.
[Github](https://github.com/RichardErkhov)
[Discord](https://discord.gg/pvy7H8DZMG)
[Request more models](https://github.com/RichardErkhov/quant_request)
sqlcoder-7b-2 - bnb 8bits
- Model creator: https://huggingface.co/defog/
- Original model: https://huggingface.co/defog/sqlcoder-7b-2/
Original model description:
---
license: cc-by-sa-4.0
library_name: transformers
pipeline_tag: text-generation
---
# Update notice
The model weights were updated at 7 AM UTC on Feb 7, 2024. The new model weights lead to a much more performant model – particularly for joins.
If you downloaded the model before that, please redownload the weights for best performance.
# Model Card for SQLCoder-7B-2
A capable large language model for natural language to SQL generation.

## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [Defog, Inc](https://defog.ai)
- **Model type:** [Text to SQL]
- **License:** [CC-by-SA-4.0]
- **Finetuned from model:** [CodeLlama-7B]
### Model Sources [optional]
- [**HuggingFace:**](https://huggingface.co/defog/sqlcoder-70b-alpha)
- [**GitHub:**](https://github.com/defog-ai/sqlcoder)
- [**Demo:**](https://defog.ai/sqlcoder-demo/)
## Uses
This model is intended to be used by non-technical users to understand data inside their SQL databases. It is meant as an analytics tool, and not as a database admin tool.
This model has not been trained to reject malicious requests from users with write access to databases, and should only be used by users with read-only access.
## How to Get Started with the Model
Use the code [here](https://github.com/defog-ai/sqlcoder/blob/main/inference.py) to get started with the model.
## Prompt
Please use the following prompt for optimal results. Please remember to use `do_sample=False` and `num_beams=4` for optimal results.
```
### Task
Generate a SQL query to answer [QUESTION]{user_question}[/QUESTION]
### Database Schema
The query will run on a database with the following schema:
{table_metadata_string_DDL_statements}
### Answer
Given the database schema, here is the SQL query that [QUESTION]{user_question}[/QUESTION]
[SQL]
```
## Evaluation
This model was evaluated on [SQL-Eval](https://github.com/defog-ai/sql-eval), a PostgreSQL based evaluation framework developed by Defog for testing and alignment of model capabilities.
You can read more about the methodology behind SQLEval [here](https://defog.ai/blog/open-sourcing-sqleval/).
### Results
We classified each generated question into one of 6 categories. The table displays the percentage of questions answered correctly by each model, broken down by category.
|                | date | group_by | order_by | ratio | join | where |
| -------------- | ---- | -------- | -------- | ----- | ---- | ----- |
| sqlcoder-70b   | 96   | 91.4     | 97.1     | 85.7  | 97.1 | 91.4  |
| sqlcoder-7b-2  | 96   | 91.4     | 94.3     | 91.4  | 94.3 | 77.1  |
| sqlcoder-34b   | 80   | 94.3     | 85.7     | 77.1  | 85.7 | 80    |
| gpt-4          | 72   | 94.3     | 97.1     | 80    | 91.4 | 80    |
| gpt-4-turbo    | 76   | 91.4     | 91.4     | 62.8  | 88.6 | 77.1  |
| natural-sql-7b | 56   | 88.6     | 85.7     | 60    | 88.6 | 80    |
| sqlcoder-7b    | 64   | 82.9     | 74.3     | 54.3  | 74.3 | 74.3  |
| gpt-3.5        | 72   | 77.1     | 82.8     | 34.3  | 65.7 | 71.4  |
| claude-2       | 52   | 71.4     | 74.3     | 57.1  | 65.7 | 62.9  |
## Model Card Contact
Contact us on X at [@defogdata](https://twitter.com/defogdata), or on email at [[email protected]](mailto:[email protected])
 | 
	{} | 
	RichardErkhov/defog_-_sqlcoder-7b-2-8bits | null | 
	[
  "safetensors",
  "region:us"
] | null | 
	2024-05-03T19:13:29+00:00 | 
	[] | 
	[] | 
	TAGS
#safetensors #region-us 
 | 
	Quantization made by Richard Erkhov.
Github
Discord
Request more models
sqlcoder-7b-2 - bnb 8bits
* Model creator: URL
* Original model: URL
Original model description:
---------------------------
license: cc-by-sa-4.0
library\_name: transformers
pipeline\_tag: text-generation
--------------------------------------------------------------------------------
Update notice
=============
The model weights were updated at 7 AM UTC on Feb 7, 2024. The new model weights lead to a much more performant model – particularly for joins.
If you downloaded the model before that, please redownload the weights for best performance.
Model Card for SQLCoder-7B-2
============================
A capable large language model for natural language to SQL generation.
!image/png
Model Details
-------------
### Model Description
This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.
* Developed by: Defog, Inc
* Model type: [Text to SQL]
* License: [CC-by-SA-4.0]
* Finetuned from model: [CodeLlama-7B]
### Model Sources [optional]
* HuggingFace:
* GitHub:
* Demo:
Uses
----
This model is intended to be used by non-technical users to understand data inside their SQL databases. It is meant as an analytics tool, and not as a database admin tool.
This model has not been trained to reject malicious requests from users with write access to databases, and should only be used by users with read-only access.
How to Get Started with the Model
---------------------------------
Use the code here to get started with the model.
Prompt
------
Please use the following prompt for optimal results. Please remember to use 'do\_sample=False' and 'num\_beams=4' for optimal results.
Evaluation
----------
This model was evaluated on SQL-Eval, a PostgreSQL based evaluation framework developed by Defog for testing and alignment of model capabilities.
You can read more about the methodology behind SQLEval here.
### Results
We classified each generated question into one of 6 categories. The table displays the percentage of questions answered correctly by each model, broken down by category.
Model Card Contact
------------------
Contact us on X at @defogdata, or on email at founders@URL
 | 
	[
  "### Model Description\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n\n* Developed by: Defog, Inc\n* Model type: [Text to SQL]\n* License: [CC-by-SA-4.0]\n* Finetuned from model: [CodeLlama-7B]",
  "### Model Sources [optional]\n\n\n* HuggingFace:\n* GitHub:\n* Demo:\n\n\nUses\n----\n\n\nThis model is intended to be used by non-technical users to understand data inside their SQL databases. It is meant as an analytics tool, and not as a database admin tool.\n\n\nThis model has not been trained to reject malicious requests from users with write access to databases, and should only be used by users with read-only access.\n\n\nHow to Get Started with the Model\n---------------------------------\n\n\nUse the code here to get started with the model.\n\n\nPrompt\n------\n\n\nPlease use the following prompt for optimal results. Please remember to use 'do\\_sample=False' and 'num\\_beams=4' for optimal results.\n\n\nEvaluation\n----------\n\n\nThis model was evaluated on SQL-Eval, a PostgreSQL based evaluation framework developed by Defog for testing and alignment of model capabilities.\n\n\nYou can read more about the methodology behind SQLEval here.",
  "### Results\n\n\nWe classified each generated question into one of 6 categories. The table displays the percentage of questions answered correctly by each model, broken down by category.\n\n\n\nModel Card Contact\n------------------\n\n\nContact us on X at @defogdata, or on email at founders@URL"
] | 
	[
  "TAGS\n#safetensors #region-us \n",
  "### Model Description\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n\n* Developed by: Defog, Inc\n* Model type: [Text to SQL]\n* License: [CC-by-SA-4.0]\n* Finetuned from model: [CodeLlama-7B]",
  "### Model Sources [optional]\n\n\n* HuggingFace:\n* GitHub:\n* Demo:\n\n\nUses\n----\n\n\nThis model is intended to be used by non-technical users to understand data inside their SQL databases. It is meant as an analytics tool, and not as a database admin tool.\n\n\nThis model has not been trained to reject malicious requests from users with write access to databases, and should only be used by users with read-only access.\n\n\nHow to Get Started with the Model\n---------------------------------\n\n\nUse the code here to get started with the model.\n\n\nPrompt\n------\n\n\nPlease use the following prompt for optimal results. Please remember to use 'do\\_sample=False' and 'num\\_beams=4' for optimal results.\n\n\nEvaluation\n----------\n\n\nThis model was evaluated on SQL-Eval, a PostgreSQL based evaluation framework developed by Defog for testing and alignment of model capabilities.\n\n\nYou can read more about the methodology behind SQLEval here.",
  "### Results\n\n\nWe classified each generated question into one of 6 categories. The table displays the percentage of questions answered correctly by each model, broken down by category.\n\n\n\nModel Card Contact\n------------------\n\n\nContact us on X at @defogdata, or on email at founders@URL"
] | 
	[
  9,
  76,
  241,
  73
] | 
	[
  "TAGS\n#safetensors #region-us \n### Model Description\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n\n* Developed by: Defog, Inc\n* Model type: [Text to SQL]\n* License: [CC-by-SA-4.0]\n* Finetuned from model: [CodeLlama-7B]### Model Sources [optional]\n\n\n* HuggingFace:\n* GitHub:\n* Demo:\n\n\nUses\n----\n\n\nThis model is intended to be used by non-technical users to understand data inside their SQL databases. It is meant as an analytics tool, and not as a database admin tool.\n\n\nThis model has not been trained to reject malicious requests from users with write access to databases, and should only be used by users with read-only access.\n\n\nHow to Get Started with the Model\n---------------------------------\n\n\nUse the code here to get started with the model.\n\n\nPrompt\n------\n\n\nPlease use the following prompt for optimal results. Please remember to use 'do\\_sample=False' and 'num\\_beams=4' for optimal results.\n\n\nEvaluation\n----------\n\n\nThis model was evaluated on SQL-Eval, a PostgreSQL based evaluation framework developed by Defog for testing and alignment of model capabilities.\n\n\nYou can read more about the methodology behind SQLEval here.### Results\n\n\nWe classified each generated question into one of 6 categories. The table displays the percentage of questions answered correctly by each model, broken down by category.\n\n\n\nModel Card Contact\n------------------\n\n\nContact us on X at @defogdata, or on email at founders@URL"
] | 
| null | null | 
	
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# GUE_tf_2-seqsight_4096_512_15M-L8_f
This model is a fine-tuned version of [mahdibaghbanzadeh/seqsight_4096_512_15M](https://huggingface.co/mahdibaghbanzadeh/seqsight_4096_512_15M) on the [mahdibaghbanzadeh/GUE_tf_2](https://huggingface.co/datasets/mahdibaghbanzadeh/GUE_tf_2) dataset.
It achieves the following results on the evaluation set:
- Loss: 0.4762
- F1 Score: 0.7889
- Accuracy: 0.789
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0005
- train_batch_size: 128
- eval_batch_size: 128
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- training_steps: 10000
### Training results
| Training Loss | Epoch | Step  | Validation Loss | F1 Score | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|:--------:|
| 0.5955        | 1.34  | 200   | 0.5400          | 0.7222   | 0.723    |
| 0.5387        | 2.68  | 400   | 0.5279          | 0.7378   | 0.738    |
| 0.5259        | 4.03  | 600   | 0.5212          | 0.7440   | 0.744    |
| 0.5216        | 5.37  | 800   | 0.5194          | 0.7418   | 0.742    |
| 0.5186        | 6.71  | 1000  | 0.5178          | 0.7380   | 0.738    |
| 0.5121        | 8.05  | 1200  | 0.5113          | 0.7370   | 0.737    |
| 0.5072        | 9.4   | 1400  | 0.5088          | 0.7378   | 0.738    |
| 0.5049        | 10.74 | 1600  | 0.5100          | 0.7390   | 0.739    |
| 0.5029        | 12.08 | 1800  | 0.5164          | 0.7475   | 0.748    |
| 0.4997        | 13.42 | 2000  | 0.5137          | 0.7435   | 0.744    |
| 0.5004        | 14.77 | 2200  | 0.5058          | 0.7422   | 0.743    |
| 0.4932        | 16.11 | 2400  | 0.5088          | 0.7445   | 0.745    |
| 0.4954        | 17.45 | 2600  | 0.5046          | 0.7419   | 0.742    |
| 0.489         | 18.79 | 2800  | 0.4987          | 0.7417   | 0.742    |
| 0.4875        | 20.13 | 3000  | 0.5027          | 0.7400   | 0.74     |
| 0.486         | 21.48 | 3200  | 0.5136          | 0.7389   | 0.74     |
| 0.4861        | 22.82 | 3400  | 0.5056          | 0.7339   | 0.734    |
| 0.4817        | 24.16 | 3600  | 0.4967          | 0.7400   | 0.74     |
| 0.4779        | 25.5  | 3800  | 0.4973          | 0.7370   | 0.737    |
| 0.4792        | 26.85 | 4000  | 0.5002          | 0.7398   | 0.74     |
| 0.4759        | 28.19 | 4200  | 0.5024          | 0.7369   | 0.737    |
| 0.4746        | 29.53 | 4400  | 0.5073          | 0.7470   | 0.747    |
| 0.4749        | 30.87 | 4600  | 0.5034          | 0.7409   | 0.741    |
| 0.4733        | 32.21 | 4800  | 0.4998          | 0.7419   | 0.742    |
| 0.4726        | 33.56 | 5000  | 0.5061          | 0.7393   | 0.74     |
| 0.4737        | 34.9  | 5200  | 0.5063          | 0.7414   | 0.742    |
| 0.4669        | 36.24 | 5400  | 0.4962          | 0.7449   | 0.745    |
| 0.469         | 37.58 | 5600  | 0.5000          | 0.7450   | 0.745    |
| 0.4658        | 38.93 | 5800  | 0.5001          | 0.7380   | 0.738    |
| 0.4631        | 40.27 | 6000  | 0.5003          | 0.7379   | 0.738    |
| 0.464         | 41.61 | 6200  | 0.4970          | 0.7400   | 0.74     |
| 0.4623        | 42.95 | 6400  | 0.5046          | 0.7459   | 0.746    |
| 0.46          | 44.3  | 6600  | 0.5083          | 0.7489   | 0.749    |
| 0.4634        | 45.64 | 6800  | 0.5060          | 0.7437   | 0.744    |
| 0.4588        | 46.98 | 7000  | 0.5045          | 0.7439   | 0.744    |
| 0.4597        | 48.32 | 7200  | 0.5028          | 0.746    | 0.746    |
| 0.4557        | 49.66 | 7400  | 0.5030          | 0.7510   | 0.751    |
| 0.4585        | 51.01 | 7600  | 0.5068          | 0.7386   | 0.739    |
| 0.4579        | 52.35 | 7800  | 0.5012          | 0.7440   | 0.744    |
| 0.4594        | 53.69 | 8000  | 0.5003          | 0.7460   | 0.746    |
| 0.4561        | 55.03 | 8200  | 0.5002          | 0.7450   | 0.745    |
| 0.4584        | 56.38 | 8400  | 0.5024          | 0.7428   | 0.743    |
| 0.4565        | 57.72 | 8600  | 0.5004          | 0.7470   | 0.747    |
| 0.4528        | 59.06 | 8800  | 0.5026          | 0.7459   | 0.746    |
| 0.4547        | 60.4  | 9000  | 0.5034          | 0.7458   | 0.746    |
| 0.4547        | 61.74 | 9200  | 0.5012          | 0.7459   | 0.746    |
| 0.4584        | 63.09 | 9400  | 0.5009          | 0.7459   | 0.746    |
| 0.4507        | 64.43 | 9600  | 0.5012          | 0.7489   | 0.749    |
| 0.4539        | 65.77 | 9800  | 0.5020          | 0.7469   | 0.747    |
| 0.4504        | 67.11 | 10000 | 0.5006          | 0.7470   | 0.747    |
### Framework versions
- PEFT 0.9.0
- Transformers 4.38.2
- Pytorch 2.2.0+cu121
- Datasets 2.17.1
- Tokenizers 0.15.2 | 
	{"library_name": "peft", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "mahdibaghbanzadeh/seqsight_4096_512_15M", "model-index": [{"name": "GUE_tf_2-seqsight_4096_512_15M-L8_f", "results": []}]} | 
	mahdibaghbanzadeh/GUE_tf_2-seqsight_4096_512_15M-L8_f | null | 
	[
  "region:us"
] | null | 
	2024-05-03T19:14:29+00:00 | 
	[] | 
	[] | 
	TAGS
#region-us 
 | 
	GUE\_tf\_2-seqsight\_4096\_512\_15M-L8\_f
=========================================
This model is a fine-tuned version of mahdibaghbanzadeh/seqsight\_4096\_512\_15M on the mahdibaghbanzadeh/GUE\_tf\_2 dataset.
It achieves the following results on the evaluation set:
* Loss: 0.4762
* F1 Score: 0.7889
* Accuracy: 0.789
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0005
* train\_batch\_size: 128
* eval\_batch\_size: 128
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* training\_steps: 10000
### Training results
### Framework versions
* PEFT 0.9.0
* Transformers 4.38.2
* Pytorch 2.2.0+cu121
* Datasets 2.17.1
* Tokenizers 0.15.2
 | 
	[
  "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000",
  "### Training results",
  "### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
	[
  "TAGS\n#region-us \n",
  "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000",
  "### Training results",
  "### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
	[
  5,
  100,
  5,
  52
] | 
	[
  "TAGS\n#region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000### Training results### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
| null | null | 
	
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# GUE_tf_2-seqsight_4096_512_15M-L32_f
This model is a fine-tuned version of [mahdibaghbanzadeh/seqsight_4096_512_15M](https://huggingface.co/mahdibaghbanzadeh/seqsight_4096_512_15M) on the [mahdibaghbanzadeh/GUE_tf_2](https://huggingface.co/datasets/mahdibaghbanzadeh/GUE_tf_2) dataset.
It achieves the following results on the evaluation set:
- Loss: 0.4798
- F1 Score: 0.7869
- Accuracy: 0.787
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0005
- train_batch_size: 128
- eval_batch_size: 128
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- training_steps: 10000
### Training results
| Training Loss | Epoch | Step  | Validation Loss | F1 Score | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|:--------:|
| 0.578         | 1.34  | 200   | 0.5342          | 0.7303   | 0.733    |
| 0.531         | 2.68  | 400   | 0.5274          | 0.7439   | 0.745    |
| 0.5177        | 4.03  | 600   | 0.5157          | 0.7400   | 0.74     |
| 0.5099        | 5.37  | 800   | 0.5128          | 0.7489   | 0.749    |
| 0.5048        | 6.71  | 1000  | 0.5149          | 0.7448   | 0.745    |
| 0.4968        | 8.05  | 1200  | 0.5041          | 0.7375   | 0.738    |
| 0.4897        | 9.4   | 1400  | 0.5042          | 0.7520   | 0.752    |
| 0.486         | 10.74 | 1600  | 0.5024          | 0.7480   | 0.748    |
| 0.4817        | 12.08 | 1800  | 0.5059          | 0.7574   | 0.758    |
| 0.4755        | 13.42 | 2000  | 0.5121          | 0.7437   | 0.744    |
| 0.4763        | 14.77 | 2200  | 0.5078          | 0.7339   | 0.736    |
| 0.4663        | 16.11 | 2400  | 0.5129          | 0.7577   | 0.758    |
| 0.4685        | 17.45 | 2600  | 0.5037          | 0.7478   | 0.748    |
| 0.4603        | 18.79 | 2800  | 0.4975          | 0.7444   | 0.745    |
| 0.4557        | 20.13 | 3000  | 0.5109          | 0.7469   | 0.747    |
| 0.4502        | 21.48 | 3200  | 0.5222          | 0.7300   | 0.731    |
| 0.4525        | 22.82 | 3400  | 0.5181          | 0.7539   | 0.754    |
| 0.4457        | 24.16 | 3600  | 0.5046          | 0.7480   | 0.748    |
| 0.4382        | 25.5  | 3800  | 0.5103          | 0.7479   | 0.748    |
| 0.4378        | 26.85 | 4000  | 0.5076          | 0.7479   | 0.748    |
| 0.4323        | 28.19 | 4200  | 0.5127          | 0.7404   | 0.741    |
| 0.4281        | 29.53 | 4400  | 0.5187          | 0.7369   | 0.737    |
| 0.4288        | 30.87 | 4600  | 0.5104          | 0.7460   | 0.746    |
| 0.4232        | 32.21 | 4800  | 0.5187          | 0.7560   | 0.756    |
| 0.4203        | 33.56 | 5000  | 0.5202          | 0.7537   | 0.754    |
| 0.4205        | 34.9  | 5200  | 0.5271          | 0.7454   | 0.746    |
| 0.409         | 36.24 | 5400  | 0.5216          | 0.7489   | 0.749    |
| 0.4114        | 37.58 | 5600  | 0.5241          | 0.7477   | 0.748    |
| 0.4077        | 38.93 | 5800  | 0.5173          | 0.7479   | 0.748    |
| 0.404         | 40.27 | 6000  | 0.5202          | 0.7560   | 0.756    |
| 0.4026        | 41.61 | 6200  | 0.5207          | 0.7430   | 0.743    |
| 0.3983        | 42.95 | 6400  | 0.5391          | 0.7477   | 0.748    |
| 0.3954        | 44.3  | 6600  | 0.5431          | 0.7377   | 0.738    |
| 0.3973        | 45.64 | 6800  | 0.5416          | 0.7351   | 0.736    |
| 0.3911        | 46.98 | 7000  | 0.5404          | 0.7419   | 0.742    |
| 0.3916        | 48.32 | 7200  | 0.5340          | 0.7429   | 0.743    |
| 0.3874        | 49.66 | 7400  | 0.5330          | 0.7450   | 0.745    |
| 0.3831        | 51.01 | 7600  | 0.5419          | 0.7387   | 0.739    |
| 0.3811        | 52.35 | 7800  | 0.5460          | 0.7430   | 0.743    |
| 0.3823        | 53.69 | 8000  | 0.5400          | 0.7440   | 0.744    |
| 0.3795        | 55.03 | 8200  | 0.5479          | 0.7407   | 0.741    |
| 0.3828        | 56.38 | 8400  | 0.5518          | 0.7407   | 0.741    |
| 0.379         | 57.72 | 8600  | 0.5405          | 0.7458   | 0.746    |
| 0.3751        | 59.06 | 8800  | 0.5438          | 0.7388   | 0.739    |
| 0.3759        | 60.4  | 9000  | 0.5491          | 0.7407   | 0.741    |
| 0.3729        | 61.74 | 9200  | 0.5489          | 0.7458   | 0.746    |
| 0.3759        | 63.09 | 9400  | 0.5501          | 0.7437   | 0.744    |
| 0.3732        | 64.43 | 9600  | 0.5483          | 0.7388   | 0.739    |
| 0.375         | 65.77 | 9800  | 0.5503          | 0.7446   | 0.745    |
| 0.369         | 67.11 | 10000 | 0.5488          | 0.7369   | 0.737    |
### Framework versions
- PEFT 0.9.0
- Transformers 4.38.2
- Pytorch 2.2.0+cu121
- Datasets 2.17.1
- Tokenizers 0.15.2 | 
	{"library_name": "peft", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "mahdibaghbanzadeh/seqsight_4096_512_15M", "model-index": [{"name": "GUE_tf_2-seqsight_4096_512_15M-L32_f", "results": []}]} | 
	mahdibaghbanzadeh/GUE_tf_2-seqsight_4096_512_15M-L32_f | null | 
	[
  "region:us"
] | null | 
	2024-05-03T19:15:29+00:00 | 
	[] | 
	[] | 
	TAGS
#region-us 
 | 
	GUE\_tf\_2-seqsight\_4096\_512\_15M-L32\_f
==========================================
This model is a fine-tuned version of mahdibaghbanzadeh/seqsight\_4096\_512\_15M on the mahdibaghbanzadeh/GUE\_tf\_2 dataset.
It achieves the following results on the evaluation set:
* Loss: 0.4798
* F1 Score: 0.7869
* Accuracy: 0.787
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0005
* train\_batch\_size: 128
* eval\_batch\_size: 128
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* training\_steps: 10000
### Training results
### Framework versions
* PEFT 0.9.0
* Transformers 4.38.2
* Pytorch 2.2.0+cu121
* Datasets 2.17.1
* Tokenizers 0.15.2
 | 
	[
  "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000",
  "### Training results",
  "### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
	[
  "TAGS\n#region-us \n",
  "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000",
  "### Training results",
  "### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
	[
  5,
  100,
  5,
  52
] | 
	[
  "TAGS\n#region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000### Training results### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
| null | null | 
	
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# GUE_virus_covid-seqsight_4096_512_15M-L1_f
This model is a fine-tuned version of [mahdibaghbanzadeh/seqsight_4096_512_15M](https://huggingface.co/mahdibaghbanzadeh/seqsight_4096_512_15M) on the [mahdibaghbanzadeh/GUE_virus_covid](https://huggingface.co/datasets/mahdibaghbanzadeh/GUE_virus_covid) dataset.
It achieves the following results on the evaluation set:
- Loss: 1.9282
- F1 Score: 0.2779
- Accuracy: 0.2832
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0005
- train_batch_size: 128
- eval_batch_size: 128
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- training_steps: 10000
### Training results
| Training Loss | Epoch | Step  | Validation Loss | F1 Score | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|:--------:|
| 2.1857        | 0.35  | 200   | 2.1860          | 0.0716   | 0.1278   |
| 2.183         | 0.7   | 400   | 2.1840          | 0.0572   | 0.1247   |
| 2.1793        | 1.05  | 600   | 2.1784          | 0.0832   | 0.1352   |
| 2.1765        | 1.4   | 800   | 2.1736          | 0.0908   | 0.1396   |
| 2.1694        | 1.75  | 1000  | 2.1685          | 0.1077   | 0.1492   |
| 2.1674        | 2.09  | 1200  | 2.1739          | 0.0989   | 0.1451   |
| 2.1644        | 2.44  | 1400  | 2.1706          | 0.1217   | 0.1480   |
| 2.1612        | 2.79  | 1600  | 2.1591          | 0.1324   | 0.1615   |
| 2.1568        | 3.14  | 1800  | 2.1538          | 0.1270   | 0.1687   |
| 2.1522        | 3.49  | 2000  | 2.1578          | 0.1333   | 0.1689   |
| 2.1535        | 3.84  | 2200  | 2.1464          | 0.1515   | 0.1814   |
| 2.1446        | 4.19  | 2400  | 2.1428          | 0.1506   | 0.1716   |
| 2.1418        | 4.54  | 2600  | 2.1369          | 0.1560   | 0.1859   |
| 2.138         | 4.89  | 2800  | 2.1304          | 0.1727   | 0.1887   |
| 2.133         | 5.24  | 3000  | 2.1352          | 0.1610   | 0.1908   |
| 2.1303        | 5.58  | 3200  | 2.1227          | 0.1787   | 0.2040   |
| 2.127         | 5.93  | 3400  | 2.1300          | 0.1451   | 0.1809   |
| 2.121         | 6.28  | 3600  | 2.1118          | 0.1827   | 0.2046   |
| 2.1132        | 6.63  | 3800  | 2.0989          | 0.1781   | 0.2027   |
| 2.11          | 6.98  | 4000  | 2.0828          | 0.2078   | 0.2254   |
| 2.0955        | 7.33  | 4200  | 2.0556          | 0.2196   | 0.2338   |
| 2.0834        | 7.68  | 4400  | 2.0488          | 0.2224   | 0.2342   |
| 2.0747        | 8.03  | 4600  | 2.0685          | 0.1803   | 0.2083   |
| 2.0662        | 8.38  | 4800  | 2.0344          | 0.2150   | 0.2323   |
| 2.0627        | 8.73  | 5000  | 2.0267          | 0.2107   | 0.2333   |
| 2.0541        | 9.08  | 5200  | 2.0213          | 0.2244   | 0.2355   |
| 2.0482        | 9.42  | 5400  | 2.0056          | 0.2347   | 0.2490   |
| 2.0413        | 9.77  | 5600  | 2.0041          | 0.2293   | 0.2441   |
| 2.0395        | 10.12 | 5800  | 1.9909          | 0.2505   | 0.2573   |
| 2.0322        | 10.47 | 6000  | 1.9841          | 0.2563   | 0.2616   |
| 2.0275        | 10.82 | 6200  | 1.9875          | 0.2414   | 0.2515   |
| 2.0227        | 11.17 | 6400  | 1.9840          | 0.2401   | 0.2509   |
| 2.0205        | 11.52 | 6600  | 1.9861          | 0.2374   | 0.2514   |
| 2.0191        | 11.87 | 6800  | 1.9717          | 0.2484   | 0.2594   |
| 2.0118        | 12.22 | 7000  | 1.9615          | 0.2657   | 0.2700   |
| 2.008         | 12.57 | 7200  | 1.9528          | 0.2658   | 0.2708   |
| 2.0108        | 12.91 | 7400  | 1.9626          | 0.2555   | 0.2638   |
| 2.0043        | 13.26 | 7600  | 1.9508          | 0.2567   | 0.2681   |
| 1.9972        | 13.61 | 7800  | 1.9566          | 0.2538   | 0.2635   |
| 1.9999        | 13.96 | 8000  | 1.9473          | 0.2719   | 0.2755   |
| 1.9947        | 14.31 | 8200  | 1.9432          | 0.2678   | 0.2758   |
| 1.9987        | 14.66 | 8400  | 1.9337          | 0.2747   | 0.2785   |
| 1.9902        | 15.01 | 8600  | 1.9422          | 0.2650   | 0.2717   |
| 1.9921        | 15.36 | 8800  | 1.9332          | 0.2762   | 0.2783   |
| 1.9841        | 15.71 | 9000  | 1.9405          | 0.2699   | 0.2780   |
| 1.9876        | 16.06 | 9200  | 1.9298          | 0.2772   | 0.2806   |
| 1.9878        | 16.4  | 9400  | 1.9299          | 0.2749   | 0.2798   |
| 1.9869        | 16.75 | 9600  | 1.9348          | 0.2755   | 0.2804   |
| 1.9865        | 17.1  | 9800  | 1.9314          | 0.2739   | 0.2793   |
| 1.9921        | 17.45 | 10000 | 1.9304          | 0.2764   | 0.2804   |
### Framework versions
- PEFT 0.9.0
- Transformers 4.38.2
- Pytorch 2.2.0+cu121
- Datasets 2.17.1
- Tokenizers 0.15.2 | 
	{"library_name": "peft", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "mahdibaghbanzadeh/seqsight_4096_512_15M", "model-index": [{"name": "GUE_virus_covid-seqsight_4096_512_15M-L1_f", "results": []}]} | 
	mahdibaghbanzadeh/GUE_virus_covid-seqsight_4096_512_15M-L1_f | null | 
	[
  "region:us"
] | null | 
	2024-05-03T19:16:42+00:00 | 
	[] | 
	[] | 
	TAGS
#region-us 
 | 
	GUE\_virus\_covid-seqsight\_4096\_512\_15M-L1\_f
================================================
This model is a fine-tuned version of mahdibaghbanzadeh/seqsight\_4096\_512\_15M on the mahdibaghbanzadeh/GUE\_virus\_covid dataset.
It achieves the following results on the evaluation set:
* Loss: 1.9282
* F1 Score: 0.2779
* Accuracy: 0.2832
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0005
* train\_batch\_size: 128
* eval\_batch\_size: 128
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* training\_steps: 10000
### Training results
### Framework versions
* PEFT 0.9.0
* Transformers 4.38.2
* Pytorch 2.2.0+cu121
* Datasets 2.17.1
* Tokenizers 0.15.2
 | 
	[
  "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000",
  "### Training results",
  "### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
	[
  "TAGS\n#region-us \n",
  "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000",
  "### Training results",
  "### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
	[
  5,
  100,
  5,
  52
] | 
	[
  "TAGS\n#region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000### Training results### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
| null | null | 
	
# Model Card for Model ID
<!-- Provide a quick summary of what the model is/does. -->
## Model Details
### Model Description
<!-- Provide a longer summary of what this model is. -->
This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
- **Developed by:** [More Information Needed]
- **Funded by [optional]:** [More Information Needed]
- **Shared by [optional]:** [More Information Needed]
- **Model type:** [More Information Needed]
- **Language(s) (NLP):** [More Information Needed]
- **License:** [More Information Needed]
- **Finetuned from model [optional]:** [More Information Needed]
### Model Sources [optional]
<!-- Provide the basic links for the model. -->
- **Repository:** [More Information Needed]
- **Paper [optional]:** [More Information Needed]
- **Demo [optional]:** [More Information Needed]
## Uses
<!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
### Direct Use
<!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
[More Information Needed]
### Downstream Use [optional]
<!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
[More Information Needed]
### Out-of-Scope Use
<!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
[More Information Needed]
## Bias, Risks, and Limitations
<!-- This section is meant to convey both technical and sociotechnical limitations. -->
[More Information Needed]
### Recommendations
<!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
[More Information Needed]
## Training Details
### Training Data
<!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
[More Information Needed]
### Training Procedure 
<!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
#### Preprocessing [optional]
[More Information Needed]
#### Training Hyperparameters
- **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
#### Speeds, Sizes, Times [optional]
<!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
[More Information Needed]
## Evaluation
<!-- This section describes the evaluation protocols and provides the results. -->
### Testing Data, Factors & Metrics
#### Testing Data
<!-- This should link to a Dataset Card if possible. -->
[More Information Needed]
#### Factors
<!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
[More Information Needed]
#### Metrics
<!-- These are the evaluation metrics being used, ideally with a description of why. -->
[More Information Needed]
### Results
[More Information Needed]
#### Summary
## Model Examination [optional]
<!-- Relevant interpretability work for the model goes here -->
[More Information Needed]
## Environmental Impact
<!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
- **Hardware Type:** [More Information Needed]
- **Hours used:** [More Information Needed]
- **Cloud Provider:** [More Information Needed]
- **Compute Region:** [More Information Needed]
- **Carbon Emitted:** [More Information Needed]
## Technical Specifications [optional]
### Model Architecture and Objective
[More Information Needed]
### Compute Infrastructure
[More Information Needed]
#### Hardware
[More Information Needed]
#### Software
[More Information Needed]
## Citation [optional]
<!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
**BibTeX:**
[More Information Needed]
**APA:**
[More Information Needed]
## Glossary [optional]
<!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
[More Information Needed]
## More Information [optional]
[More Information Needed]
## Model Card Authors [optional]
[More Information Needed]
## Model Card Contact
[More Information Needed]
 | 
	{"library_name": "transformers", "tags": []} | 
	golf2248/ofn2ele | null | 
	[
  "region:us"
] | null | 
	2024-05-03T19:16:49+00:00 | 
	[] | 
	[] | 
	TAGS
#region-us 
 | 
	
# Model Card for Model ID
## Model Details
### Model Description
This is the model card of a  transformers model that has been pushed on the Hub. This model card has been automatically generated.
- Developed by: 
- Funded by [optional]: 
- Shared by [optional]: 
- Model type: 
- Language(s) (NLP): 
- License: 
- Finetuned from model [optional]: 
### Model Sources [optional]
- Repository: 
- Paper [optional]: 
- Demo [optional]: 
## Uses
### Direct Use
### Downstream Use [optional]
### Out-of-Scope Use
## Bias, Risks, and Limitations
### Recommendations
Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
## How to Get Started with the Model
Use the code below to get started with the model.
## Training Details
### Training Data
### Training Procedure 
#### Preprocessing [optional]
#### Training Hyperparameters
- Training regime:  
#### Speeds, Sizes, Times [optional]
## Evaluation
### Testing Data, Factors & Metrics
#### Testing Data
#### Factors
#### Metrics
### Results
#### Summary
## Model Examination [optional]
## Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).
- Hardware Type: 
- Hours used: 
- Cloud Provider: 
- Compute Region: 
- Carbon Emitted: 
## Technical Specifications [optional]
### Model Architecture and Objective
### Compute Infrastructure
#### Hardware
#### Software
[optional]
BibTeX:
APA:
## Glossary [optional]
## More Information [optional]
## Model Card Authors [optional]
## Model Card Contact
 | 
	[
  "# Model Card for Model ID",
  "## Model Details",
  "### Model Description\n\n\n\nThis is the model card of a  transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
  "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
  "## Uses",
  "### Direct Use",
  "### Downstream Use [optional]",
  "### Out-of-Scope Use",
  "## Bias, Risks, and Limitations",
  "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
  "## How to Get Started with the Model\n\nUse the code below to get started with the model.",
  "## Training Details",
  "### Training Data",
  "### Training Procedure",
  "#### Preprocessing [optional]",
  "#### Training Hyperparameters\n\n- Training regime:",
  "#### Speeds, Sizes, Times [optional]",
  "## Evaluation",
  "### Testing Data, Factors & Metrics",
  "#### Testing Data",
  "#### Factors",
  "#### Metrics",
  "### Results",
  "#### Summary",
  "## Model Examination [optional]",
  "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
  "## Technical Specifications [optional]",
  "### Model Architecture and Objective",
  "### Compute Infrastructure",
  "#### Hardware",
  "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
  "## Glossary [optional]",
  "## More Information [optional]",
  "## Model Card Authors [optional]",
  "## Model Card Contact"
] | 
	[
  "TAGS\n#region-us \n",
  "# Model Card for Model ID",
  "## Model Details",
  "### Model Description\n\n\n\nThis is the model card of a  transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:",
  "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:",
  "## Uses",
  "### Direct Use",
  "### Downstream Use [optional]",
  "### Out-of-Scope Use",
  "## Bias, Risks, and Limitations",
  "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.",
  "## How to Get Started with the Model\n\nUse the code below to get started with the model.",
  "## Training Details",
  "### Training Data",
  "### Training Procedure",
  "#### Preprocessing [optional]",
  "#### Training Hyperparameters\n\n- Training regime:",
  "#### Speeds, Sizes, Times [optional]",
  "## Evaluation",
  "### Testing Data, Factors & Metrics",
  "#### Testing Data",
  "#### Factors",
  "#### Metrics",
  "### Results",
  "#### Summary",
  "## Model Examination [optional]",
  "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:",
  "## Technical Specifications [optional]",
  "### Model Architecture and Objective",
  "### Compute Infrastructure",
  "#### Hardware",
  "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:",
  "## Glossary [optional]",
  "## More Information [optional]",
  "## Model Card Authors [optional]",
  "## Model Card Contact"
] | 
	[
  5,
  6,
  4,
  75,
  23,
  3,
  5,
  8,
  9,
  8,
  34,
  20,
  4,
  5,
  5,
  11,
  13,
  12,
  3,
  10,
  6,
  5,
  6,
  4,
  5,
  7,
  49,
  7,
  7,
  5,
  5,
  15,
  7,
  7,
  8,
  5
] | 
	[
  "TAGS\n#region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a  transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact"
] | 
| null | null | 
	
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
should probably proofread and complete it, then remove this comment. -->
# GUE_virus_covid-seqsight_4096_512_15M-L8_f
This model is a fine-tuned version of [mahdibaghbanzadeh/seqsight_4096_512_15M](https://huggingface.co/mahdibaghbanzadeh/seqsight_4096_512_15M) on the [mahdibaghbanzadeh/GUE_virus_covid](https://huggingface.co/datasets/mahdibaghbanzadeh/GUE_virus_covid) dataset.
It achieves the following results on the evaluation set:
- Loss: 1.5376
- F1 Score: 0.4255
- Accuracy: 0.4204
## Model description
More information needed
## Intended uses & limitations
More information needed
## Training and evaluation data
More information needed
## Training procedure
### Training hyperparameters
The following hyperparameters were used during training:
- learning_rate: 0.0005
- train_batch_size: 128
- eval_batch_size: 128
- seed: 42
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
- lr_scheduler_type: linear
- training_steps: 10000
### Training results
| Training Loss | Epoch | Step  | Validation Loss | F1 Score | Accuracy |
|:-------------:|:-----:|:-----:|:---------------:|:--------:|:--------:|
| 2.1853        | 0.35  | 200   | 2.1837          | 0.0874   | 0.1342   |
| 2.1805        | 0.7   | 400   | 2.1771          | 0.0925   | 0.1295   |
| 2.1715        | 1.05  | 600   | 2.1691          | 0.1098   | 0.1369   |
| 2.1617        | 1.4   | 800   | 2.1614          | 0.1110   | 0.1502   |
| 2.1494        | 1.75  | 1000  | 2.1453          | 0.1504   | 0.1804   |
| 2.1399        | 2.09  | 1200  | 2.1424          | 0.1226   | 0.1747   |
| 2.11          | 2.44  | 1400  | 2.0699          | 0.1890   | 0.2135   |
| 2.0694        | 2.79  | 1600  | 2.0231          | 0.2130   | 0.2406   |
| 2.0288        | 3.14  | 1800  | 2.0134          | 0.2062   | 0.2318   |
| 1.9962        | 3.49  | 2000  | 1.9502          | 0.2475   | 0.2598   |
| 1.9712        | 3.84  | 2200  | 1.8961          | 0.2710   | 0.2816   |
| 1.9382        | 4.19  | 2400  | 1.8577          | 0.2936   | 0.2901   |
| 1.9121        | 4.54  | 2600  | 1.8328          | 0.3132   | 0.3178   |
| 1.8976        | 4.89  | 2800  | 1.8175          | 0.3134   | 0.3129   |
| 1.875         | 5.24  | 3000  | 1.7826          | 0.3280   | 0.3340   |
| 1.8617        | 5.58  | 3200  | 1.7518          | 0.3499   | 0.3488   |
| 1.8365        | 5.93  | 3400  | 1.7553          | 0.3296   | 0.3388   |
| 1.8209        | 6.28  | 3600  | 1.7260          | 0.3515   | 0.3516   |
| 1.8059        | 6.63  | 3800  | 1.7081          | 0.3620   | 0.3599   |
| 1.8003        | 6.98  | 4000  | 1.7012          | 0.3732   | 0.3702   |
| 1.7834        | 7.33  | 4200  | 1.6943          | 0.3664   | 0.3658   |
| 1.7706        | 7.68  | 4400  | 1.6790          | 0.3783   | 0.3660   |
| 1.767         | 8.03  | 4600  | 1.6793          | 0.3684   | 0.3688   |
| 1.7547        | 8.38  | 4800  | 1.6680          | 0.3748   | 0.3752   |
| 1.7509        | 8.73  | 5000  | 1.6592          | 0.3763   | 0.3802   |
| 1.7496        | 9.08  | 5200  | 1.6561          | 0.3869   | 0.3803   |
| 1.7273        | 9.42  | 5400  | 1.6421          | 0.3869   | 0.3880   |
| 1.7283        | 9.77  | 5600  | 1.6331          | 0.3979   | 0.3955   |
| 1.725         | 10.12 | 5800  | 1.6186          | 0.4024   | 0.3932   |
| 1.7221        | 10.47 | 6000  | 1.6145          | 0.3986   | 0.3946   |
| 1.7101        | 10.82 | 6200  | 1.6078          | 0.4082   | 0.4012   |
| 1.6922        | 11.17 | 6400  | 1.6023          | 0.4073   | 0.4024   |
| 1.6973        | 11.52 | 6600  | 1.5917          | 0.4116   | 0.4045   |
| 1.6989        | 11.87 | 6800  | 1.5862          | 0.4106   | 0.4053   |
| 1.684         | 12.22 | 7000  | 1.5780          | 0.4176   | 0.4108   |
| 1.674         | 12.57 | 7200  | 1.5750          | 0.4172   | 0.4123   |
| 1.6799        | 12.91 | 7400  | 1.5693          | 0.4194   | 0.4140   |
| 1.6687        | 13.26 | 7600  | 1.5574          | 0.4183   | 0.4153   |
| 1.6716        | 13.61 | 7800  | 1.5663          | 0.4222   | 0.4162   |
| 1.6615        | 13.96 | 8000  | 1.5567          | 0.4226   | 0.4177   |
| 1.6562        | 14.31 | 8200  | 1.5533          | 0.4217   | 0.4166   |
| 1.6584        | 14.66 | 8400  | 1.5481          | 0.4290   | 0.4196   |
| 1.656         | 15.01 | 8600  | 1.5455          | 0.4272   | 0.4237   |
| 1.6563        | 15.36 | 8800  | 1.5480          | 0.4297   | 0.4204   |
| 1.639         | 15.71 | 9000  | 1.5463          | 0.4260   | 0.4224   |
| 1.6507        | 16.06 | 9200  | 1.5438          | 0.4242   | 0.4192   |
| 1.6477        | 16.4  | 9400  | 1.5385          | 0.4275   | 0.4226   |
| 1.6475        | 16.75 | 9600  | 1.5404          | 0.4289   | 0.4243   |
| 1.6414        | 17.1  | 9800  | 1.5406          | 0.4294   | 0.4249   |
| 1.6511        | 17.45 | 10000 | 1.5388          | 0.4300   | 0.4249   |
### Framework versions
- PEFT 0.9.0
- Transformers 4.38.2
- Pytorch 2.2.0+cu121
- Datasets 2.17.1
- Tokenizers 0.15.2 | 
	{"library_name": "peft", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "base_model": "mahdibaghbanzadeh/seqsight_4096_512_15M", "model-index": [{"name": "GUE_virus_covid-seqsight_4096_512_15M-L8_f", "results": []}]} | 
	mahdibaghbanzadeh/GUE_virus_covid-seqsight_4096_512_15M-L8_f | null | 
	[
  "region:us"
] | null | 
	2024-05-03T19:16:59+00:00 | 
	[] | 
	[] | 
	TAGS
#region-us 
 | 
	GUE\_virus\_covid-seqsight\_4096\_512\_15M-L8\_f
================================================
This model is a fine-tuned version of mahdibaghbanzadeh/seqsight\_4096\_512\_15M on the mahdibaghbanzadeh/GUE\_virus\_covid dataset.
It achieves the following results on the evaluation set:
* Loss: 1.5376
* F1 Score: 0.4255
* Accuracy: 0.4204
Model description
-----------------
More information needed
Intended uses & limitations
---------------------------
More information needed
Training and evaluation data
----------------------------
More information needed
Training procedure
------------------
### Training hyperparameters
The following hyperparameters were used during training:
* learning\_rate: 0.0005
* train\_batch\_size: 128
* eval\_batch\_size: 128
* seed: 42
* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
* lr\_scheduler\_type: linear
* training\_steps: 10000
### Training results
### Framework versions
* PEFT 0.9.0
* Transformers 4.38.2
* Pytorch 2.2.0+cu121
* Datasets 2.17.1
* Tokenizers 0.15.2
 | 
	[
  "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000",
  "### Training results",
  "### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
	[
  "TAGS\n#region-us \n",
  "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000",
  "### Training results",
  "### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
	[
  5,
  100,
  5,
  52
] | 
	[
  "TAGS\n#region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0005\n* train\\_batch\\_size: 128\n* eval\\_batch\\_size: 128\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* training\\_steps: 10000### Training results### Framework versions\n\n\n* PEFT 0.9.0\n* Transformers 4.38.2\n* Pytorch 2.2.0+cu121\n* Datasets 2.17.1\n* Tokenizers 0.15.2"
] | 
			Subsets and Splits
				
	
				
			
				
No community queries yet
The top public SQL queries from the community will appear here once available.