sha
null
last_modified
null
library_name
stringclasses
154 values
text
stringlengths
1
900k
metadata
stringlengths
2
348k
pipeline_tag
stringclasses
45 values
id
stringlengths
5
122
tags
sequencelengths
1
1.84k
created_at
stringlengths
25
25
arxiv
sequencelengths
0
201
languages
sequencelengths
0
1.83k
tags_str
stringlengths
17
9.34k
text_str
stringlengths
0
389k
text_lists
sequencelengths
0
722
processed_texts
sequencelengths
1
723
tokens_length
sequencelengths
1
723
input_texts
sequencelengths
1
61
embeddings
sequencelengths
768
768
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> This modelcard aims to be a base template for new models. It has been generated using [this raw template](https://github.com/huggingface/huggingface_hub/blob/main/src/huggingface_hub/templates/modelcard_template.md?plain=1). ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"language": ["en"], "license": "apache-2.0", "tags": ["spam", "ham", "text-classification", "spam-detection", "transformer", "BERT"], "metrics": ["accuracy", "precision", "recall", "f1"]}
text-classification
surajkarki/bert_spam_detection
[ "transformers", "safetensors", "bert", "text-classification", "spam", "ham", "spam-detection", "transformer", "BERT", "en", "arxiv:1910.09700", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-10T11:41:04+00:00
[ "1910.09700" ]
[ "en" ]
TAGS #transformers #safetensors #bert #text-classification #spam #ham #spam-detection #transformer #BERT #en #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# Model Card for Model ID This modelcard aims to be a base template for new models. It has been generated using this raw template. ## Model Details ### Model Description - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID\n\n\n\nThis modelcard aims to be a base template for new models. It has been generated using this raw template.", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #bert #text-classification #spam #ham #spam-detection #transformer #BERT #en #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# Model Card for Model ID\n\n\n\nThis modelcard aims to be a base template for new models. It has been generated using this raw template.", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 73, 29, 3, 54, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #bert #text-classification #spam #ham #spam-detection #transformer #BERT #en #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for Model ID\n\n\n\nThis modelcard aims to be a base template for new models. It has been generated using this raw template.## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.060846053063869476, 0.23102636635303497, -0.005311095155775547, 0.018827879801392555, 0.09810111671686172, 0.006204136647284031, 0.06435392796993256, 0.11026522517204285, -0.044758230447769165, 0.11905689537525177, 0.04496736079454422, 0.10052280873060226, 0.11213523894548416, 0.14894649386405945, -0.011842279694974422, -0.2270294427871704, 0.042261771857738495, -0.09480778872966766, -0.01131212618201971, 0.12356702983379364, 0.15160435438156128, -0.10081950575113297, 0.07366161793470383, -0.021857889369130135, -0.02774098329246044, -0.02955561876296997, -0.05628713220357895, -0.046249162405729294, 0.038911692798137665, 0.04460589960217476, 0.07276396453380585, -0.0012976675061509013, 0.09512961655855179, -0.29318609833717346, 0.013180354610085487, 0.05916278064250946, -0.0010204463033005595, 0.07329036295413971, 0.08824049681425095, -0.06161319836974144, 0.0990297719836235, -0.05787641555070877, 0.13778537511825562, 0.07876460254192352, -0.09176216274499893, -0.1657382994890213, -0.08194044977426529, 0.11996110528707504, 0.1786220669746399, 0.0583060160279274, -0.03338391333818436, 0.10778108239173889, -0.08241261541843414, 0.018075326457619667, 0.051230140030384064, -0.11803103983402252, -0.061731647700071335, 0.08014608174562454, 0.09924096614122391, 0.051868733018636703, -0.12278079241514206, -0.026057250797748566, 0.02093029022216797, 0.02498387172818184, 0.08117860555648804, 0.015459233894944191, 0.1428638994693756, 0.0331612266600132, -0.13725554943084717, -0.06425867974758148, 0.13114507496356964, 0.044054754078388214, -0.03782825544476509, -0.2226717472076416, -0.014341924339532852, -0.016292652115225792, -0.038734935224056244, -0.04603470489382744, 0.04724094271659851, 0.00258267973549664, 0.09334701299667358, -0.011935672722756863, -0.07319381088018417, -0.036301080137491226, 0.08941294997930527, 0.037524912506341934, 0.023960456252098083, -0.02138778194785118, 0.0185980387032032, 0.10872919112443924, 0.08089233189821243, -0.11515794694423676, -0.04210402071475983, -0.05640871450304985, -0.05811947584152222, -0.043289124965667725, 0.03443212807178497, 0.034827712923288345, 0.07910986244678497, 0.24691930413246155, 0.030251773074269295, 0.04802825301885605, 0.03419145941734314, 0.00985373742878437, 0.04771697148680687, 0.0988336056470871, -0.061134032905101776, -0.14369793236255646, -0.019393252208828926, 0.09684329479932785, 0.0044943238608539104, -0.02413606271147728, -0.035790108144283295, 0.05557147413492203, 0.05313596501946449, 0.1087687537074089, 0.0901719331741333, 0.02194131910800934, -0.07592466473579407, -0.04747409000992775, 0.19265827536582947, -0.15763290226459503, 0.04181480407714844, 0.03595896065235138, -0.029345853254199028, -0.051915764808654785, 0.011654418893158436, 0.0405878908932209, -0.03962789848446846, 0.09083618223667145, -0.05960635468363762, -0.05014261603355408, -0.10636767745018005, -0.034399162977933884, 0.042562324553728104, 0.01723947562277317, -0.029428275302052498, -0.025811292231082916, -0.09159603714942932, -0.08556561172008514, 0.09069214016199112, -0.06303957104682922, -0.08029936254024506, -0.03705522045493126, -0.07685603201389313, 0.021398086100816727, 0.017544128000736237, 0.08259475976228714, -0.02697903662919998, 0.0430714413523674, -0.05945727601647377, 0.03915499895811081, 0.09799584746360779, 0.03359437733888626, -0.07025362551212311, 0.07407977432012558, -0.20500430464744568, 0.09351687878370285, -0.08237084746360779, 0.05404892563819885, -0.16517356038093567, -0.021646773442626, 0.035842251032590866, 0.012746496126055717, -0.0023608608171343803, 0.13809989392757416, -0.1951623558998108, -0.015793023630976677, 0.17055751383304596, -0.09860983490943909, -0.0813264325261116, 0.05483976751565933, -0.05858084186911583, 0.11603755503892899, 0.043703872710466385, 0.027368007227778435, 0.05482223629951477, -0.10091445595026016, -0.01074003241956234, -0.04694300517439842, -0.0005621294840238988, 0.11841494590044022, 0.07941167801618576, -0.09625284373760223, 0.01453226339071989, 0.019353436306118965, -0.05617493391036987, -0.06352175027132034, -0.03164803236722946, -0.103378526866436, 0.011011202819645405, -0.08043006807565689, 0.015635645017027855, -0.010524747893214226, -0.09148982167243958, -0.028604550287127495, -0.15444032847881317, -0.0035872836597263813, 0.07542518526315689, -0.006032976787537336, -0.022344162687659264, -0.10161427408456802, 0.028328029438853264, 0.004334146156907082, -0.015752505511045456, -0.12836691737174988, -0.03311048820614815, 0.029864473268389702, -0.1398722231388092, 0.024762729182839394, -0.06979484856128693, 0.05152558162808418, 0.01594889909029007, -0.03684169426560402, -0.02313959412276745, 0.018041841685771942, 0.0181815717369318, -0.03663019835948944, -0.22595837712287903, -0.029148021712899208, -0.03657178953289986, 0.1614484190940857, -0.23333902657032013, 0.04273020103573799, 0.0592685341835022, 0.14000216126441956, -0.001800570054911077, -0.05503767356276512, 0.02805313467979431, -0.06213228777050972, -0.01852474734187126, -0.05650646612048149, 0.003787156892940402, -0.011114315129816532, -0.05211605876684189, 0.016202395781874657, -0.15842702984809875, -0.030281051993370056, 0.09450609982013702, 0.05496407300233841, -0.12023564428091049, -0.037088628858327866, -0.02692631632089615, -0.05368422344326973, -0.044952500611543655, -0.06030070409178734, 0.10378193855285645, 0.05644598975777626, 0.04264364019036293, -0.07165750116109848, -0.07193780690431595, -0.008732444606721401, -0.0283542200922966, -0.027244621887803078, 0.09791266173124313, 0.07764574885368347, -0.12314607948064804, 0.09824379533529282, 0.08174357563257217, 0.05268559604883194, 0.0808856338262558, -0.01777013950049877, -0.06946681439876556, -0.02488362044095993, 0.03438793867826462, 0.017969751730561256, 0.1247405856847763, -0.07361793518066406, 0.043318621814250946, 0.04215524345636368, -0.03371783345937729, 0.024378051981329918, -0.08279484510421753, 0.015950947999954224, 0.02009696699678898, -0.026583930477499962, 0.01819591224193573, -0.04391464591026306, 0.014110244810581207, 0.08433371037244797, 0.04992838203907013, 0.019030826166272163, 0.021681327372789383, -0.05090608075261116, -0.11987829953432083, 0.16361065208911896, -0.11325320601463318, -0.20004428923130035, -0.13761267066001892, 0.030820457264780998, 0.028881702572107315, -0.015643291175365448, -0.00011878925579367206, -0.05725417286157608, -0.10149629414081573, -0.09856011718511581, 0.0001100537774618715, 0.05015863478183746, -0.0914674699306488, -0.044806379824876785, 0.032500606030225754, 0.05414127930998802, -0.13855455815792084, 0.017858097329735756, 0.047405730932950974, -0.07920131087303162, -0.021346740424633026, 0.056585922837257385, 0.08935081213712692, 0.19479824602603912, 0.01166493259370327, -0.008502036333084106, 0.030966736376285553, 0.24817651510238647, -0.1290121227502823, 0.10277467221021652, 0.1186867505311966, -0.07627835124731064, 0.08415593206882477, 0.2184683233499527, 0.04561953246593475, -0.09162671118974686, 0.02285509556531906, 0.04396403580904007, -0.01536225713789463, -0.2476857751607895, -0.0740656927227974, -0.0015460761496797204, -0.0774923637509346, 0.09010779112577438, 0.09123805165290833, 0.08705777674913406, 0.041413191705942154, -0.08066001534461975, -0.08635467290878296, 0.06439684331417084, 0.10551072657108307, -0.01280728168785572, 0.006937025114893913, 0.08600689470767975, -0.03055952675640583, 0.012527638114988804, 0.08392724394798279, 0.0179476086050272, 0.1414586901664734, 0.0550127737224102, 0.15687808394432068, 0.0900038555264473, 0.08691256493330002, 0.002381687518209219, 0.021418960765004158, 0.007541146129369736, 0.046011898666620255, 0.00012496292765717953, -0.08136370778083801, -0.011426543816924095, 0.11158295720815659, 0.045678991824388504, 0.02691865898668766, 0.013544561341404915, -0.054519884288311005, 0.07280656695365906, 0.21316269040107727, 0.0037951769772917032, -0.20765647292137146, -0.05136663094162941, 0.08109947293996811, -0.09554574638605118, -0.11038394272327423, -0.0036544781178236008, 0.023280145600438118, -0.172376349568367, 0.04734580218791962, -0.03552211821079254, 0.11291784048080444, -0.09887423366308212, -0.019115284085273743, 0.07343356311321259, 0.056466709822416306, -0.02035517804324627, 0.07093282788991928, -0.1843399554491043, 0.11618699133396149, 0.010180775076150894, 0.0679173618555069, -0.09027552604675293, 0.0861331969499588, 0.00296687800437212, -0.00016919136396609247, 0.15839342772960663, -0.001057981513440609, -0.06550402194261551, -0.06754417717456818, -0.09252142906188965, -0.00001871532185759861, 0.08856557309627533, -0.13567925989627838, 0.0825386792421341, -0.028672153130173683, -0.030781608074903488, -0.0009386496967636049, -0.09815806895494461, -0.13395968079566956, -0.16980260610580444, 0.04287174716591835, -0.11449810862541199, 0.04835556074976921, -0.10311508178710938, -0.031101534143090248, -0.03617340698838234, 0.17408965528011322, -0.21066473424434662, -0.0837070643901825, -0.13421645760536194, -0.09898757934570312, 0.14019335806369781, -0.042919013649225235, 0.09299955517053604, -0.007974348030984402, 0.16059836745262146, 0.011771610006690025, -0.009377176873385906, 0.07654373347759247, -0.08686360716819763, -0.19723525643348694, -0.06356918811798096, 0.16022488474845886, 0.11800024658441544, 0.035040467977523804, 0.006772933062165976, 0.0348261259496212, -0.02509423904120922, -0.10873053967952728, 0.012905186042189598, 0.14028729498386383, 0.0864093229174614, 0.008932836353778839, -0.029537536203861237, -0.12568023800849915, -0.08583091199398041, -0.030917931348085403, 0.027260128408670425, 0.1652447134256363, -0.07574866712093353, 0.15030373632907867, 0.1313382238149643, -0.07285097986459732, -0.2004021257162094, 0.00821303203701973, 0.027626486495137215, -0.006531629245728254, 0.01970193162560463, -0.17472802102565765, 0.09417473524808884, 0.01364482194185257, -0.05699317157268524, 0.07889273017644882, -0.18617285788059235, -0.13536621630191803, 0.09020738303661346, 0.05016392469406128, -0.20890682935714722, -0.13661623001098633, -0.09721614420413971, -0.04583297297358513, -0.14623665809631348, 0.08007040619850159, 0.008560382761061192, 0.0002896931837312877, 0.035596903413534164, 0.013956155627965927, 0.027758408337831497, -0.05222964659333229, 0.1867751032114029, 0.009660015814006329, 0.034506916999816895, -0.0841507837176323, -0.10261005908250809, 0.03125438094139099, -0.04989323392510414, 0.08211404830217361, -0.038368530571460724, 0.01546419132500887, -0.11711467057466507, -0.0406205840408802, -0.05830693989992142, 0.02281222492456436, -0.0988258495926857, -0.0864490494132042, -0.04663041606545448, 0.08264776319265366, 0.10456380993127823, -0.016543271020054817, -0.03031996451318264, -0.0871538519859314, 0.06987126171588898, 0.21347269415855408, 0.19843274354934692, 0.07269161939620972, -0.06584971398115158, 0.00524877430871129, -0.026451846584677696, 0.049611177295446396, -0.19544798135757446, 0.05965208262205124, 0.059413421899080276, 0.016592655330896378, 0.11072653532028198, -0.026375140994787216, -0.14233891665935516, -0.07295814156532288, 0.06251874566078186, -0.06316804140806198, -0.2033347189426422, 0.003745071357116103, 0.04412570223212242, -0.18055826425552368, -0.041150711476802826, 0.04130204766988754, -0.020838387310504913, -0.036795683205127716, 0.0079710828140378, 0.0912938266992569, -0.013863525353372097, 0.09111254662275314, 0.08112890273332596, 0.08930916339159012, -0.09691885858774185, 0.08404511213302612, 0.09860573709011078, -0.04531831294298172, 0.03026443161070347, 0.08709584176540375, -0.05332306772470474, -0.03925600275397301, 0.057474005967378616, 0.11398152261972427, 0.029085304588079453, -0.05483565852046013, 0.0021846252493560314, -0.08811620622873306, 0.05884631723165512, 0.11351381242275238, 0.024913636967539787, 0.010103213600814342, 0.050888024270534515, 0.022944912314414978, -0.07950517535209656, 0.1197190210223198, 0.06495686620473862, 0.013746139593422413, -0.050343628972768784, -0.012972835451364517, 0.007580206263810396, -0.03410564363002777, -0.0036759504582732916, -0.007042183540761471, -0.07837948948144913, -0.010691107250750065, -0.14199258387088776, 0.019668277353048325, -0.07862839102745056, 0.01932922750711441, 0.021822281181812286, -0.026770085096359253, 0.0018197429599240422, -0.0022175556514412165, -0.07307286560535431, -0.061841800808906555, -0.004609233699738979, 0.10428155213594437, -0.16260771453380585, 0.020957274362444878, 0.08871085196733475, -0.10316695272922516, 0.08952585607767105, -0.01228699367493391, -0.010681744664907455, 0.009501860477030277, -0.1604069471359253, 0.04706793278455734, -0.02419787459075451, 0.001514054019935429, 0.007158221676945686, -0.17727579176425934, -0.0031901246402412653, -0.03214691951870918, -0.06250643730163574, -0.003958316519856453, -0.0031856121495366096, -0.11072640120983124, 0.09363716840744019, 0.017669128254055977, -0.08554013073444366, -0.024503635242581367, 0.03864124417304993, 0.09648603945970535, -0.039326298981904984, 0.1473175287246704, -0.019024677574634552, 0.06165226548910141, -0.17273084819316864, -0.00936288945376873, -0.013303357176482677, 0.019549472257494926, -0.06358615309000015, -0.006228203885257244, 0.04673456400632858, -0.017971651628613472, 0.18822260200977325, -0.02065960317850113, 0.01296722237020731, 0.059390291571617126, 0.03316885605454445, 0.006605167407542467, 0.09621282666921616, 0.07412262260913849, 0.0074822986498475075, 0.003864201018586755, 0.019348157569766045, -0.04423025622963905, -0.03812021017074585, -0.17505823075771332, 0.05469284951686859, 0.20465557277202606, 0.08967997133731842, -0.014918910339474678, 0.0660412609577179, -0.10784000903367996, -0.09154842793941498, 0.13061891496181488, -0.02189461700618267, 0.0004088451969437301, -0.07496452331542969, 0.14138704538345337, 0.14295504987239838, -0.1989862322807312, 0.07120868563652039, -0.0709717720746994, -0.05314210429787636, -0.11150199174880981, -0.21307113766670227, -0.06736505031585693, -0.03906109556555748, -0.012415767647325993, -0.06315860152244568, 0.05894939601421356, 0.09261933714151382, -0.012423107400536537, -0.012606928125023842, 0.07046116888523102, -0.04066010192036629, -0.0029021534137427807, 0.032159578055143356, 0.05652230978012085, 0.020245857536792755, -0.06832777708768845, 0.011479213833808899, -0.007051642052829266, 0.05238058418035507, 0.06530535966157913, 0.03687490522861481, -0.0239978339523077, 0.01698048785328865, -0.031366799026727676, -0.1070348471403122, 0.04849597439169884, -0.020538674667477608, -0.05100303515791893, 0.14451071619987488, 0.029358869418501854, -0.0008598314598202705, -0.009032728150486946, 0.2226794809103012, -0.06903144717216492, -0.10183659195899963, -0.15399087965488434, 0.07952781766653061, -0.049747858196496964, 0.05045871436595917, 0.04802615940570831, -0.11165600270032883, 0.03645265847444534, 0.1487404853105545, 0.17148733139038086, -0.041777901351451874, 0.011177603155374527, 0.019479960203170776, 0.002867528237402439, -0.023461658507585526, 0.03911631181836128, 0.04590704292058945, 0.12784627079963684, -0.06543675810098648, 0.08279953896999359, 0.0030772718600928783, -0.07675863802433014, -0.03228495270013809, 0.12843884527683258, -0.018319766968488693, -0.0026149547193199396, -0.05678294226527214, 0.13558794558048248, -0.06799177080392838, -0.2013210505247116, 0.035109203308820724, -0.070358507335186, -0.13598571717739105, -0.0243891179561615, 0.030710391700267792, 0.0020158186089247465, 0.01866556517779827, 0.0727710872888565, -0.0669262707233429, 0.17890121042728424, 0.03497213125228882, -0.0631909891963005, -0.06223292648792267, 0.08126397430896759, -0.08107732981443405, 0.31241410970687866, 0.027101291343569756, 0.03918704390525818, 0.1060868352651596, -0.021173862740397453, -0.12597346305847168, 0.026093050837516785, 0.09948436170816422, -0.0614771693944931, 0.056627754122018814, 0.1452733725309372, -0.004123810213059187, 0.13825909793376923, 0.06740131229162216, -0.07552407681941986, 0.04708049073815346, -0.07852347195148468, -0.0732496827840805, -0.10962823778390884, 0.09456362575292587, -0.09013107419013977, 0.1466786414384842, 0.12905706465244293, -0.06086337938904762, 0.012822799384593964, -0.02333816885948181, 0.05946286767721176, -0.004934367258101702, 0.11317452788352966, 0.014301356859505177, -0.18075986206531525, 0.03012043423950672, -0.03197002038359642, 0.09713516384363174, -0.18791241943836212, -0.07808645069599152, 0.03713325783610344, 0.0012116004945710301, -0.07539481669664383, 0.11805769056081772, 0.07434976100921631, 0.02524297498166561, -0.04627189412713051, -0.029403584077954292, -0.010600801557302475, 0.14996948838233948, -0.09596016258001328, -0.005618612747639418 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text-classification
sekhharr/hackathon_v9
[ "transformers", "safetensors", "bert", "text-classification", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-10T11:43:06+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #bert #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #bert #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 46, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #bert #text-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.06817419826984406, 0.1699906885623932, -0.003845146857202053, 0.018365124240517616, 0.11478200554847717, 0.00763329304754734, 0.07986336201429367, 0.10738246887922287, -0.0269484706223011, 0.1267213374376297, 0.03862300142645836, 0.1017010435461998, 0.11044707149267197, 0.18616852164268494, 0.002953584771603346, -0.2117370218038559, 0.062315817922353745, -0.11355884373188019, 0.01421935111284256, 0.12174045294523239, 0.14285145699977875, -0.10472407191991806, 0.07340893894433975, -0.03533155843615532, -0.019184017553925514, -0.029508300125598907, -0.06138347089290619, -0.062117863446474075, 0.06899366527795792, 0.06911981105804443, 0.06776530295610428, 0.02535320073366165, 0.07980640977621078, -0.2927248775959015, 0.019224179908633232, 0.07704847306013107, 0.004596637096256018, 0.06310366839170456, 0.07900875061750412, -0.06604467332363129, 0.12630145251750946, -0.0469624362885952, 0.15577000379562378, 0.07483451068401337, -0.09700790792703629, -0.1833430528640747, -0.07868417352437973, 0.08138132095336914, 0.1542958915233612, 0.0575118213891983, -0.03566069155931473, 0.14360417425632477, -0.0863327905535698, 0.015191552229225636, 0.06608161330223083, -0.07603584229946136, -0.05265629291534424, 0.04255614057183266, 0.07708034664392471, 0.09375373274087906, -0.1291297972202301, -0.010211804881691933, 0.04229271039366722, 0.01873886212706566, 0.10347303748130798, 0.02310175821185112, 0.11163661628961563, 0.026270611211657524, -0.13941870629787445, -0.06378244608640671, 0.1267201453447342, 0.02999917045235634, -0.05697820335626602, -0.23340454697608948, -0.007031846325844526, -0.028088124468922615, -0.024382783100008965, -0.03983099386096001, 0.03844287618994713, -0.0294374767690897, 0.07875318825244904, 0.011917876079678535, -0.07096433639526367, -0.04893866181373596, 0.08819517493247986, 0.06123629957437515, 0.022971229627728462, -0.02526908740401268, 0.02413375861942768, 0.11652170121669769, 0.09283795207738876, -0.11929406225681305, -0.06425759196281433, -0.06432286649942398, -0.08888134360313416, -0.04847237840294838, 0.03574979677796364, 0.0754702165722847, 0.04938753694295883, 0.19765597581863403, 0.006366121117025614, 0.05646394565701485, 0.0260426327586174, 0.015338202007114887, 0.06355882436037064, 0.07606974244117737, -0.0483609177172184, -0.13532373309135437, -0.041331104934215546, 0.11784996092319489, 0.007102925330400467, -0.032494835555553436, -0.03608081117272377, 0.06173410639166832, 0.05820438638329506, 0.1192656010389328, 0.06626396626234055, 0.019241811707615852, -0.06749388575553894, -0.03806937485933304, 0.1874811202287674, -0.1540532261133194, 0.020778683945536613, 0.01720726117491722, -0.05474008247256279, -0.043989501893520355, 0.0171356238424778, 0.008756347931921482, -0.02707439661026001, 0.10765543580055237, -0.0681026354432106, -0.03794260695576668, -0.10775765031576157, -0.057500679045915604, 0.032596319913864136, -0.011795170605182648, -0.030085675418376923, -0.0443500280380249, -0.1081358790397644, -0.07622874528169632, 0.06656987965106964, -0.06241556629538536, -0.07165607810020447, -0.03565853461623192, -0.05456356331706047, 0.012712954543530941, 0.002376573858782649, 0.12743701040744781, -0.02916865609586239, 0.04608776792883873, -0.04567936435341835, 0.06814887374639511, 0.13260088860988617, 0.03273140639066696, -0.07753180712461472, 0.0658058449625969, -0.21566881239414215, 0.10687019675970078, -0.09710393846035004, 0.030530039221048355, -0.1602926403284073, -0.027380328625440598, 0.025517668575048447, 0.035233598202466965, -0.01142354216426611, 0.1405038684606552, -0.18839864432811737, -0.036833859980106354, 0.17594264447689056, -0.13455410301685333, -0.09238629788160324, 0.06278568506240845, -0.057844966650009155, 0.12792403995990753, 0.05209182947874069, -0.027332304045557976, 0.059202857315540314, -0.13285812735557556, -0.024411480873823166, -0.0557100772857666, -0.0024997375439852476, 0.1512058526277542, 0.06197551265358925, -0.05537422001361847, 0.02062765136361122, 0.020016051828861237, -0.024297641590237617, -0.045233841985464096, -0.034582652151584625, -0.0977277010679245, 0.006374812684953213, -0.07783913612365723, 0.015467152930796146, -0.014978265389800072, -0.08572793006896973, -0.037934768944978714, -0.15898989140987396, -0.0011305080261081457, 0.09650373458862305, 0.007345336955040693, -0.029424650594592094, -0.09241348505020142, 0.005526319146156311, 0.014208783395588398, -0.01407501008361578, -0.15675009787082672, -0.05031281337141991, 0.03119790367782116, -0.16866113245487213, 0.033627450466156006, -0.04903757572174072, 0.03549545630812645, 0.04459671676158905, -0.04535774141550064, -0.02160848118364811, 0.0152364457026124, 0.017460787668824196, -0.02394135482609272, -0.24046528339385986, -0.016492176800966263, -0.049182213842868805, 0.17930001020431519, -0.24510087072849274, 0.04199686273932457, 0.062341514974832535, 0.12092601507902145, 0.005246761720627546, -0.047405339777469635, 0.03611646965146065, -0.04782456159591675, -0.04614211246371269, -0.06458985060453415, -0.004041698761284351, -0.03005247749388218, -0.04619463160634041, 0.04105473682284355, -0.19605930149555206, -0.029964644461870193, 0.11028317362070084, 0.07146124541759491, -0.1701718270778656, -0.07740049809217453, -0.03032514825463295, -0.06061795726418495, -0.09144899994134903, -0.04754206910729408, 0.10501570999622345, 0.0424359068274498, 0.054926108568906784, -0.07243066281080246, -0.047703035175800323, 0.012159520760178566, -0.008316845633089542, -0.035265736281871796, 0.0910128578543663, 0.09147894382476807, -0.1183665320277214, 0.1003284826874733, 0.06719938665628433, 0.061502620577812195, 0.10171586275100708, 0.005867301486432552, -0.09559345990419388, -0.012123096734285355, 0.023821083828806877, 0.014739413745701313, 0.13627171516418457, -0.08041682839393616, 0.03041158802807331, 0.043761420994997025, -0.03445654734969139, 0.011279189959168434, -0.10341424494981766, 0.02347799763083458, 0.03186830133199692, -0.007050554268062115, 0.025736309587955475, -0.054652560502290726, 0.013161799870431423, 0.1042112186551094, 0.03211836516857147, 0.0227707140147686, 0.015011876821517944, -0.03876445069909096, -0.12403564900159836, 0.17888623476028442, -0.09523385018110275, -0.25718894600868225, -0.12982366979122162, 0.0025806569028645754, 0.04723223298788071, -0.01322246715426445, 0.01721704937517643, -0.057064954191446304, -0.10620168596506119, -0.10562704503536224, 0.017637979239225388, 0.05363597348332405, -0.08985256403684616, -0.06360358744859695, 0.05353172495961189, 0.038684699684381485, -0.12286891043186188, 0.023170825093984604, 0.04556644707918167, -0.0685787945985794, 0.004107215907424688, 0.05788148567080498, 0.08483386784791946, 0.18220773339271545, 0.013182112947106361, -0.017085859552025795, 0.012520790100097656, 0.22458304464817047, -0.14599265158176422, 0.09336943179368973, 0.13670575618743896, -0.0603153258562088, 0.08385994285345078, 0.20927630364894867, 0.031639765948057175, -0.09247095137834549, 0.04077373072504997, 0.032938770949840546, -0.040111273527145386, -0.23512989282608032, -0.07784179598093033, 0.0005755177116952837, -0.07578593492507935, 0.10564399510622025, 0.09113350510597229, 0.11394096910953522, 0.05373004451394081, -0.10628228634595871, -0.06785868853330612, 0.04576247185468674, 0.11892180144786835, -0.020387137308716774, 0.0034232554025948048, 0.09533460438251495, -0.032669007778167725, 0.016892950981855392, 0.0903218612074852, 0.010076770558953285, 0.18146716058254242, 0.040793538093566895, 0.12895575165748596, 0.08216089755296707, 0.06404399126768112, 0.023877892643213272, 0.01690720207989216, 0.028041476383805275, 0.02853785827755928, -0.021422842517495155, -0.08959300816059113, -0.01811058260500431, 0.14208537340164185, 0.03174193948507309, 0.030387144535779953, 0.009561240673065186, -0.0344390794634819, 0.0656830444931984, 0.16341377794742584, 0.01373966783285141, -0.23032663762569427, -0.06265294551849365, 0.07538370788097382, -0.07251506298780441, -0.11472991853952408, -0.007447437848895788, 0.029569825157523155, -0.17949488759040833, 0.045079123228788376, -0.02245110087096691, 0.1028464064002037, -0.11004801839590073, -0.024476202204823494, 0.04228143393993378, 0.06811302900314331, -0.03619502857327461, 0.07936927676200867, -0.21071307361125946, 0.14414268732070923, 0.0071875168941915035, 0.0627245381474495, -0.10963346809148788, 0.08230046182870865, 0.02151823230087757, 0.009466269053518772, 0.16101586818695068, -0.0074920570477843285, -0.09318114817142487, -0.07651645690202713, -0.07556641101837158, -0.011319656856358051, 0.09559466689825058, -0.10184428840875626, 0.08486217260360718, -0.008358954451978207, -0.03313955292105675, -0.00388424564152956, -0.1140027567744255, -0.13622364401817322, -0.18601436913013458, 0.05523287504911423, -0.11181046068668365, 0.03691478446125984, -0.11166879534721375, -0.06252610683441162, -0.02911795862019062, 0.19807842373847961, -0.1904531568288803, -0.08140338957309723, -0.14539870619773865, -0.07204011082649231, 0.12212951481342316, -0.04274967685341835, 0.07663191109895706, 0.00015701932716183364, 0.2071707546710968, -0.004644640255719423, 0.0014644638868048787, 0.0856679305434227, -0.09557735919952393, -0.206184521317482, -0.09439684450626373, 0.13821037113666534, 0.12497473508119583, 0.04596934840083122, -0.0036321566440165043, 0.024304913356900215, -0.0027867835015058517, -0.10976199060678482, 0.02332260087132454, 0.1432444006204605, 0.08416087180376053, 0.03885705769062042, -0.02675866149365902, -0.14533737301826477, -0.1054752767086029, -0.05289754271507263, 0.019448768347501755, 0.17674845457077026, -0.07222644239664078, 0.1607094258069992, 0.15837931632995605, -0.06414622813463211, -0.20734171569347382, 0.032782182097435, 0.03679283335804939, -0.011663361452519894, 0.03244366869330406, -0.20815548300743103, 0.07330463081598282, 0.016213007271289825, -0.06075131520628929, 0.1363404095172882, -0.1705039143562317, -0.14891991019248962, 0.0919104814529419, 0.07189090549945831, -0.2193969339132309, -0.13394345343112946, -0.09907522052526474, -0.055755600333213806, -0.10410746932029724, 0.08695419132709503, 0.014253350906074047, 0.004559517838060856, 0.040003977715969086, 0.024713784456253052, 0.021094202995300293, -0.05303549766540527, 0.19554594159126282, -0.004308625590056181, 0.041122131049633026, -0.08143328875303268, -0.08729361742734909, 0.030160382390022278, -0.06146852299571037, 0.07429458200931549, -0.02577015943825245, 0.004456855356693268, -0.1102396696805954, -0.06384536623954773, -0.05289682373404503, 0.03639809414744377, -0.08915901929140091, -0.0958789587020874, -0.05767008289694786, 0.10389325767755508, 0.08919540792703629, -0.03324571251869202, -0.058615610003471375, -0.10058292001485825, 0.0726626068353653, 0.22699709236621857, 0.18807223439216614, 0.07284927368164062, -0.07015843689441681, 0.0006279588560573757, -0.022037893533706665, 0.05516184866428375, -0.20622296631336212, 0.04608523100614548, 0.042553652077913284, 0.028887338936328888, 0.13527612388134003, -0.02506665140390396, -0.1602775603532791, -0.04527048021554947, 0.06014934554696083, -0.06545355916023254, -0.1614707112312317, -0.0005388054414652288, 0.09576781094074249, -0.16179001331329346, -0.06273222714662552, 0.024773813784122467, -0.036137934774160385, -0.025756290182471275, 0.0013679420808330178, 0.08270203322172165, 0.027825508266687393, 0.11478793621063232, 0.06896458566188812, 0.11150709539651871, -0.10231363028287888, 0.08406093716621399, 0.09299708157777786, -0.10971303284168243, 0.03247435390949249, 0.07298728823661804, -0.0610542818903923, -0.03390142321586609, 0.023122351616621017, 0.08364028483629227, 0.026266440749168396, -0.0744837298989296, -0.0008558011031709611, -0.1099681630730629, 0.06663114577531815, 0.13796411454677582, 0.032853204756975174, 0.0030810926109552383, 0.04435998201370239, 0.025823330506682396, -0.09881676733493805, 0.11186433583498001, 0.03916766867041588, 0.03720828518271446, -0.04767070338129997, 0.004865953233093023, 0.041960928589105606, -0.01269921287894249, -0.016253290697932243, -0.039693526923656464, -0.06471271812915802, -0.010708925314247608, -0.15688052773475647, 0.031037067994475365, -0.07176970690488815, 0.009115522727370262, 0.018755896016955376, -0.033779606223106384, 0.0002807097043842077, 0.0073861307464540005, -0.07919271290302277, -0.03761441633105278, -0.006646361667662859, 0.10705258697271347, -0.15747743844985962, 0.008323745802044868, 0.08949586004018784, -0.12556882202625275, 0.07766758650541306, -0.007498627994209528, -0.010838181711733341, 0.01879316382110119, -0.14380721747875214, 0.06054820865392685, -0.008177737705409527, 0.006405212916433811, 0.023949483409523964, -0.20071232318878174, 0.005702852737158537, -0.04664513096213341, -0.053938448429107666, -0.00976315326988697, -0.04211960732936859, -0.11404810100793839, 0.10492629557847977, 0.0196357611566782, -0.0860515683889389, -0.018402770161628723, 0.05309472978115082, 0.10592338442802429, -0.057369641959667206, 0.1371336728334427, -0.02283608354628086, 0.05825338885188103, -0.17831756174564362, -0.016339747235178947, -0.017454219982028008, 0.012596609070897102, -0.03102201037108898, -0.008158523589372635, 0.05483707785606384, -0.015072896145284176, 0.22714339196681976, -0.021177595481276512, 0.030790245160460472, 0.06548503786325455, 0.0070373364724218845, -0.013032838702201843, 0.08790382742881775, 0.04639120027422905, 0.021969040855765343, 0.017426103353500366, 0.016819516196846962, -0.047575462609529495, -0.019116412848234177, -0.12834098935127258, 0.08396804332733154, 0.16439755260944366, 0.08264775574207306, -0.005125291179865599, 0.053218428045511246, -0.11920209228992462, -0.08098750561475754, 0.10049403458833694, -0.033211447298526764, -0.001258186181075871, -0.057700008153915405, 0.14298145473003387, 0.15607422590255737, -0.1750815361738205, 0.06616412103176117, -0.07047461718320847, -0.05687202885746956, -0.11070677638053894, -0.17143365740776062, -0.06694129854440689, -0.03149404004216194, -0.005430171266198158, -0.06143372505903244, 0.06926561146974564, 0.10244123637676239, 0.008475886657834053, 0.002354414900764823, 0.08415096998214722, -0.033749498426914215, -0.0007962242234498262, 0.04344722256064415, 0.05283457785844803, 0.021373692899942398, -0.06691429764032364, 0.0076249162666499615, 0.004598149098455906, 0.038937900215387344, 0.05476561188697815, 0.0317605659365654, -0.014559607952833176, 0.011871086433529854, -0.013089693151414394, -0.09815122187137604, 0.03718226030468941, -0.029980625957250595, -0.0468674972653389, 0.14802806079387665, 0.01827765442430973, 0.0034919960889965296, -0.021031659096479416, 0.23128560185432434, -0.06903756409883499, -0.0798255056142807, -0.14009471237659454, 0.15071772038936615, -0.04670744761824608, 0.05065378174185753, 0.04940982535481453, -0.10087474435567856, 0.03407741338014603, 0.14691931009292603, 0.14527682960033417, -0.02467990294098854, 0.007901503704488277, 0.011187983676791191, 0.0055741616524755955, -0.025625228881835938, 0.05354921892285347, 0.04412171617150307, 0.12145667523145676, -0.06669453531503677, 0.09297986328601837, -0.007810541894286871, -0.0844663754105568, -0.02094031497836113, 0.1328510195016861, 0.0014671299140900373, 0.02338746376335621, -0.0805477648973465, 0.11851188540458679, -0.06559251248836517, -0.25864502787590027, 0.061333827674388885, -0.06666524708271027, -0.15384668111801147, -0.018917718902230263, 0.02399173192679882, 0.00401253392919898, 0.024401430040597916, 0.06268756836652756, -0.06360985338687897, 0.14903949201107025, 0.03688151761889458, -0.07834678888320923, -0.07808853685855865, 0.07696148753166199, -0.08397532254457474, 0.3018210828304291, 0.008228152059018612, 0.04951678216457367, 0.09650786966085434, -0.03327273949980736, -0.13361208140850067, 0.04569283500313759, 0.09728528559207916, -0.06408768892288208, 0.06690182536840439, 0.19748380780220032, -0.008177485316991806, 0.12026696652173996, 0.07469146698713303, -0.08128973841667175, 0.057554539293050766, -0.07613562047481537, -0.09007242321968079, -0.09192728251218796, 0.08888110518455505, -0.060599785298109055, 0.15479759871959686, 0.13393908739089966, -0.04440179467201233, -0.001819826546125114, -0.03071022778749466, 0.05197824910283089, -0.002023093169555068, 0.1104598417878151, 0.022785736247897148, -0.19388216733932495, 0.031831543892621994, -0.014316190034151077, 0.0986877828836441, -0.2479904145002365, -0.07837841659784317, 0.0403057225048542, -0.013808837160468102, -0.05274871736764908, 0.12204353511333466, 0.052187733352184296, 0.04937480762600899, -0.05449601635336876, -0.057812657207250595, -0.00025569170247763395, 0.16358551383018494, -0.1094348207116127, -0.00204258831217885 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # audio_classification This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the minds14 dataset. It achieves the following results on the evaluation set: - Loss: 2.6430 - Accuracy: 0.0708 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 3e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:--------:| | No log | 1.0 | 15 | 2.6430 | 0.0708 | | No log | 2.0 | 30 | 2.6430 | 0.0708 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.1.2+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["minds14"], "metrics": ["accuracy"], "base_model": "facebook/wav2vec2-base", "model-index": [{"name": "audio_classification", "results": [{"task": {"type": "audio-classification", "name": "Audio Classification"}, "dataset": {"name": "minds14", "type": "minds14", "config": "en-US", "split": "train", "args": "en-US"}, "metrics": [{"type": "accuracy", "value": 0.07079646017699115, "name": "Accuracy"}]}]}]}
audio-classification
andikamandalaa/audio_classification
[ "transformers", "safetensors", "wav2vec2", "audio-classification", "generated_from_trainer", "dataset:minds14", "base_model:facebook/wav2vec2-base", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
2024-02-10T11:43:55+00:00
[]
[]
TAGS #transformers #safetensors #wav2vec2 #audio-classification #generated_from_trainer #dataset-minds14 #base_model-facebook/wav2vec2-base #license-apache-2.0 #model-index #endpoints_compatible #region-us
audio\_classification ===================== This model is a fine-tuned version of facebook/wav2vec2-base on the minds14 dataset. It achieves the following results on the evaluation set: * Loss: 2.6430 * Accuracy: 0.0708 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 3e-05 * train\_batch\_size: 32 * eval\_batch\_size: 32 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 2 ### Training results ### Framework versions * Transformers 4.37.2 * Pytorch 2.1.2+cu121 * Datasets 2.16.1 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #safetensors #wav2vec2 #audio-classification #generated_from_trainer #dataset-minds14 #base_model-facebook/wav2vec2-base #license-apache-2.0 #model-index #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1" ]
[ 73, 98, 4, 33 ]
[ "passage: TAGS\n#transformers #safetensors #wav2vec2 #audio-classification #generated_from_trainer #dataset-minds14 #base_model-facebook/wav2vec2-base #license-apache-2.0 #model-index #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 3e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.2+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1" ]
[ -0.11834293603897095, 0.11365389823913574, -0.0014671443495899439, 0.09533829241991043, 0.14046558737754822, 0.003528430825099349, 0.14492540061473846, 0.07881590723991394, -0.0836125984787941, 0.039837874472141266, 0.10059697926044464, 0.08848977833986282, 0.018113821744918823, 0.1472911536693573, -0.0771133303642273, -0.19080811738967896, 0.040149662643671036, 0.00825540255755186, -0.056458450853824615, 0.11892567574977875, 0.0896621122956276, -0.12143298238515854, 0.08299870043992996, -0.011878592893481255, -0.15859928727149963, 0.0023805282544344664, 0.030708692967891693, -0.055982720106840134, 0.1184174045920372, 0.031071051955223083, 0.10480352491140366, 0.03286273032426834, 0.08934727311134338, -0.2102634310722351, 0.01653960347175598, 0.07205632328987122, -0.005030239932239056, 0.0648004487156868, 0.0522616021335125, -0.009775001555681229, 0.034237951040267944, -0.07582144439220428, 0.033542584627866745, 0.03886830434203148, -0.14158093929290771, -0.27182555198669434, -0.09450128674507141, 0.04802819713950157, 0.10639029741287231, 0.096769317984581, -0.027922948822379112, 0.12127673625946045, -0.03900729492306709, 0.08970081806182861, 0.22774150967597961, -0.27842020988464355, -0.05161116644740105, 0.039356529712677, 0.03489867225289345, 0.08993269503116608, -0.10186673700809479, 0.004149329848587513, 0.06877114623785019, 0.029566576704382896, 0.14209093153476715, -0.02969842404127121, -0.05887643247842789, -0.02171851322054863, -0.1395920068025589, -0.03686632588505745, 0.19825221598148346, 0.09915883094072342, -0.0692707747220993, -0.03589169681072235, -0.060356780886650085, -0.1148899495601654, -0.04478621855378151, -0.003052924992516637, 0.053946953266859055, -0.014388234354555607, -0.0702907145023346, 0.009722570888698101, -0.08555780351161957, -0.09402623027563095, -0.018929148092865944, 0.13440804183483124, 0.009211532771587372, 0.008988481014966965, -0.01832331158220768, 0.07740894705057144, -0.04504422843456268, -0.14930306375026703, 0.012400374747812748, 0.02426016703248024, -0.014752767980098724, -0.04775460436940193, -0.042781732976436615, -0.046463798731565475, 0.02095908299088478, 0.12367237359285355, -0.08530918508768082, 0.04267270117998123, -0.0106390081346035, 0.04675755277276039, -0.07953473180532455, 0.11557843536138535, -0.058725569397211075, -0.08190947026014328, 0.026767417788505554, 0.08238261938095093, 0.05096506327390671, -0.008790150284767151, -0.11018027365207672, 0.02785705402493477, 0.10678096115589142, 0.012348080985248089, -0.06021149829030037, 0.04430742561817169, -0.07931280881166458, -0.00983836967498064, 0.03304888308048248, -0.07269485294818878, 0.017162274569272995, 0.020162293687462807, -0.026101307943463326, -0.05296701192855835, 0.01952281966805458, 0.053406305611133575, 0.02738230861723423, 0.08149031549692154, -0.07434853166341782, -0.011085945181548595, -0.054270438849925995, -0.10316779464483261, 0.034321870654821396, -0.05641038715839386, 0.04926082864403725, -0.10730661451816559, -0.18562696874141693, -0.004277908243238926, 0.07842427492141724, -0.00460333377122879, -0.049029555171728134, -0.046865321695804596, -0.09307198971509933, 0.013500411063432693, -0.03672965243458748, 0.08402664214372635, -0.08525028079748154, 0.09648792445659637, 0.03660124912858009, 0.06350565701723099, -0.05662873759865761, 0.052040945738554, -0.08583961427211761, 0.049400150775909424, -0.20022442936897278, 0.02734602987766266, -0.07258011400699615, 0.055381279438734055, -0.09654053300619125, -0.07853947579860687, 0.005513358861207962, 0.010001505725085735, 0.06188027933239937, 0.08969621360301971, -0.1848876029253006, -0.07775041460990906, 0.15652281045913696, -0.10810371488332748, -0.16570936143398285, 0.14169320464134216, -0.04549179598689079, 0.03847943991422653, 0.07638871669769287, 0.2047618180513382, 0.07563190162181854, -0.11757465451955795, -0.005868620704859495, -0.026236901059746742, 0.05426623299717903, -0.03997492417693138, 0.10199226438999176, 0.008621116168797016, -0.011960388161242008, -0.010090494528412819, -0.048370301723480225, 0.06225684657692909, -0.07548016309738159, -0.08140064775943756, -0.029722163453698158, -0.1267451047897339, 0.04384751617908478, 0.03808212652802467, 0.016521289944648743, -0.11106465756893158, -0.07436593621969223, 0.0636313408613205, 0.09510062634944916, -0.05870095640420914, 0.01157532911747694, -0.07634278386831284, 0.0841473713517189, -0.0953352078795433, -0.03143276646733284, -0.14781063795089722, -0.012539047747850418, -0.0004962191451340914, -0.018263326957821846, 0.0009079385199584067, 0.010240519419312477, 0.06287086755037308, 0.07710237056016922, -0.07624434679746628, -0.07342275232076645, -0.036065474152565, 0.03539085015654564, -0.08182936161756516, -0.2105427384376526, -0.026991387829184532, -0.03379979729652405, 0.13842402398586273, -0.21802040934562683, 0.020279638469219208, 0.01706908829510212, 0.08581017702817917, 0.057674165815114975, -0.023419979959726334, -0.012091884389519691, 0.0613669715821743, -0.00968741811811924, -0.061204683035612106, 0.04521937295794487, 0.024374084547162056, -0.0780462846159935, -0.03465215116739273, -0.13114532828330994, 0.23609818518161774, 0.13430963456630707, -0.06613887846469879, -0.04805389419198036, 0.03310941904783249, -0.03560139611363411, -0.02783527784049511, -0.061295442283153534, -0.012383158318698406, 0.13522595167160034, 0.0021158503368496895, 0.130329430103302, -0.10591047257184982, -0.012343572452664375, 0.0515107698738575, -0.04875154420733452, 0.012518694624304771, 0.09200122207403183, 0.04498196393251419, -0.1294342279434204, 0.156575009226799, 0.21123920381069183, -0.05568072944879532, 0.14906622469425201, -0.06609051674604416, -0.06472137570381165, -0.03295597806572914, -0.014564886689186096, -0.005618321243673563, 0.1433047503232956, -0.11427062004804611, 0.02813897654414177, 0.011298148892819881, 0.04929206520318985, -0.00517816748470068, -0.20369349420070648, -0.036647360771894455, 0.045035284012556076, -0.06226543337106705, -0.07217784970998764, -0.015796896070241928, -0.017418207600712776, 0.07526280730962753, -0.0033660854678601027, -0.08961593359708786, 0.052534956485033035, -0.013510621152818203, -0.09566161036491394, 0.1706255078315735, -0.09016555547714233, -0.14416490495204926, -0.13364164531230927, -0.0655168667435646, -0.046551432460546494, 0.032011836767196655, 0.0841841921210289, -0.06961659342050552, -0.04867182672023773, -0.09221930801868439, 0.03096100501716137, 0.03271494433283806, 0.017636986449360847, 0.06779243797063828, 0.009313526563346386, 0.11578049510717392, -0.08952368050813675, -0.003991199191659689, -0.010046628303825855, -0.014613804407417774, 0.007525382563471794, 0.03594372048974037, 0.10193774849176407, 0.1275692731142044, 0.012041682377457619, -0.002454606583341956, -0.016399851068854332, 0.2855708599090576, -0.09610125422477722, -0.0028941400814801455, 0.15831109881401062, -0.036007385700941086, 0.03372659534215927, 0.16435790061950684, 0.0568171851336956, -0.1149020791053772, 0.018919920548796654, 0.037576135247945786, -0.03486056253314018, -0.20911182463169098, -0.04300374537706375, -0.03844647482037544, -0.00768089247867465, 0.05442386493086815, 0.011600310914218426, 0.016891388222575188, 0.05288922041654587, 0.00964182149618864, 0.049245063215494156, 0.008321757428348064, 0.05038078501820564, 0.06305435299873352, 0.03956114873290062, 0.10798090696334839, -0.0486537329852581, -0.01812797226011753, 0.04019740968942642, 0.009051058441400528, 0.19159166514873505, 0.05952693894505501, 0.11675294488668442, 0.07411153614521027, 0.12936130166053772, 0.007985846139490604, 0.06650199741125107, -0.009518873877823353, -0.04858247563242912, -0.01629938930273056, -0.05143118277192116, -0.03243783116340637, 0.0320885144174099, -0.07634694874286652, 0.07172796875238419, -0.1261681169271469, 0.014602871611714363, 0.04758549481630325, 0.2628569006919861, 0.0717577338218689, -0.3190153241157532, -0.10310018062591553, 0.04337482899427414, -0.034026939421892166, -0.03647163510322571, 0.04656407609581947, 0.14271695911884308, -0.04386534541845322, 0.05598742142319679, -0.03941117227077484, 0.0908237099647522, -0.02727840468287468, 0.050410058349370956, 0.056809406727552414, 0.06459067761898041, -0.01243962999433279, 0.05101696029305458, -0.23461903631687164, 0.2940439283847809, 0.03077365830540657, 0.09517084062099457, -0.02165924198925495, -0.010158714838325977, 0.030220787972211838, 0.10633979737758636, 0.11367127299308777, 0.0023103333078324795, -0.08245154470205307, -0.18789491057395935, -0.04581001028418541, 0.025281960144639015, 0.08253740519285202, -0.02686547301709652, 0.08682768791913986, -0.03572795167565346, -0.006836650427430868, 0.08536193519830704, 0.028107037767767906, -0.1087019145488739, -0.056596916168928146, -0.03585798293352127, 0.08813458681106567, 0.04965701326727867, -0.0970165953040123, -0.08686621487140656, -0.15190808475017548, 0.09348733723163605, -0.036536503583192825, -0.014693749137222767, -0.09897254407405853, 0.01958644948899746, 0.05876161903142929, -0.05808892473578453, 0.05630487576127052, 0.028551926836371422, 0.11720293015241623, 0.0156716275960207, -0.044330962002277374, 0.1270495504140854, -0.09439713507890701, -0.1862398236989975, -0.06518864631652832, 0.14520910382270813, 0.038628242909908295, 0.04970507323741913, 0.01027011964470148, 0.022622475400567055, 0.017358114942908287, -0.06466500461101532, 0.05086841061711311, -0.011516043916344643, 0.040557894855737686, 0.021466732025146484, -0.007172880694270134, -0.028852596879005432, -0.03903688117861748, -0.04239597171545029, 0.1365751028060913, 0.31375572085380554, -0.0626479983329773, 0.028499053791165352, 0.10636002570390701, -0.04590828716754913, -0.19465139508247375, 0.05863956734538078, 0.035934124141931534, 0.002995157614350319, 0.06747833639383316, -0.1534472107887268, 0.11920373141765594, 0.07348569482564926, -0.03390870615839958, 0.0701211616396904, -0.2578738331794739, -0.10879985988140106, 0.12028877437114716, 0.14110489189624786, 0.13483396172523499, -0.13753503561019897, -0.030936123803257942, -0.07223708182573318, -0.17160898447036743, 0.10273627191781998, -0.19598548114299774, 0.10494515299797058, 0.019252033904194832, 0.051831331104040146, -0.00024083540483843535, -0.04309764876961708, 0.13086146116256714, 0.027031121775507927, 0.1305084526538849, -0.045703496783971786, 0.029799945652484894, 0.07967014610767365, -0.05471550300717354, 0.014422200620174408, -0.09615838527679443, 0.03933195024728775, -0.055800702422857285, -0.029961463063955307, -0.05940314754843712, 0.01848405785858631, -0.034664053469896317, -0.06200028210878372, -0.022809742018580437, 0.04145852476358414, 0.06294813752174377, -0.017080053687095642, 0.14940744638442993, 0.00558867072686553, 0.11892125010490417, 0.13214966654777527, 0.09967377036809921, -0.09116116166114807, -0.0444527268409729, -0.007554339244961739, -0.05702783167362213, 0.07112494856119156, -0.15510523319244385, 0.07139763236045837, 0.09243584424257278, 0.014315079897642136, 0.146983802318573, 0.05663773790001869, -0.04808151721954346, 0.008937197737395763, 0.061988942325115204, -0.15378527343273163, -0.13396185636520386, -0.022856757044792175, -0.009504631161689758, -0.10734710842370987, 0.05460332706570625, 0.13066066801548004, -0.06386681646108627, 0.017795464023947716, -0.013957303948700428, 0.024188023060560226, -0.0827762633562088, 0.1701226532459259, 0.062272340059280396, 0.06174245476722717, -0.10299414396286011, 0.09878311306238174, 0.026680678129196167, -0.08505148440599442, 0.020196789875626564, 0.006398582365363836, -0.08931107074022293, -0.044819097965955734, 0.05571950599551201, 0.1882772594690323, -0.006440080236643553, -0.07849084585905075, -0.11249686032533646, -0.13479302823543549, 0.03880428895354271, 0.1948939561843872, 0.08539889007806778, 0.01394946500658989, -0.01184663362801075, 0.0049933684058487415, -0.10366439074277878, 0.11567189544439316, 0.03303400054574013, 0.07101008296012878, -0.210621640086174, 0.08565317839384079, -0.007035230286419392, -0.0036668130196630955, -0.028316836804151535, 0.042709801346063614, -0.10734325647354126, 0.021530181169509888, -0.16240383684635162, -0.00921196211129427, -0.03381263092160225, 0.017001798376441002, -0.010900960303843021, -0.06331713497638702, -0.07432584464550018, 0.032476089894771576, -0.09916727989912033, -0.013154344633221626, 0.0476616770029068, 0.06840964406728745, -0.12026095390319824, -0.04146042838692665, 0.03005213290452957, -0.07620181143283844, 0.06814149767160416, 0.028033100068569183, 0.01556357741355896, 0.05273093655705452, -0.19793817400932312, -0.010451829992234707, 0.07684803754091263, 0.007253721356391907, 0.0490257553756237, -0.1467002034187317, -0.022293103858828545, -0.00033572345273569226, 0.03416993096470833, 0.01158037781715393, 0.0875280499458313, -0.0978255495429039, -0.014862882904708385, -0.019560979679226875, -0.04249759763479233, -0.04089554026722908, 0.002671578200533986, 0.14089761674404144, 0.004922108259052038, 0.22248299419879913, -0.08019006252288818, 0.010456553660333157, -0.20231521129608154, 0.014271284453570843, -0.012843940407037735, -0.11974538117647171, -0.13999438285827637, -0.047657258808612823, 0.04362417757511139, -0.06727424263954163, 0.12818337976932526, -0.015347611159086227, 0.03277647867798805, 0.0329759381711483, -0.064180888235569, 0.02091953344643116, 0.03404631465673447, 0.23539268970489502, 0.024581387639045715, -0.03680438548326492, 0.02892483025789261, 0.012945492751896381, 0.11558306217193604, 0.10076496750116348, 0.14312595129013062, 0.20461969077587128, -0.02598392404615879, 0.08729809522628784, 0.06174499914050102, -0.06275735795497894, -0.15430913865566254, 0.04513701796531677, -0.05225079879164696, 0.09963519871234894, -0.007896630093455315, 0.18465210497379303, 0.08843297511339188, -0.16882392764091492, 0.03162989765405655, -0.07130879163742065, -0.06356494873762131, -0.1007138267159462, -0.059045713394880295, -0.10005094110965729, -0.16696663200855255, 0.003942605573683977, -0.11898024380207062, 0.016890570521354675, 0.10160285234451294, -0.011241543106734753, -0.011038141325116158, 0.18459323048591614, -0.01634666882455349, 0.030036000534892082, 0.059466294944286346, -0.006058705970644951, -0.07289580255746841, -0.052657682448625565, -0.0881565660238266, 0.06720496714115143, -0.03462354838848114, 0.018222562968730927, -0.056228455156087875, -0.0639832392334938, 0.05310872942209244, -0.0027696494944393635, -0.12093479186296463, 0.014456639066338539, 0.013516468927264214, 0.04415352642536163, 0.056136418133974075, 0.02227756194770336, 0.020853698253631592, 0.0232357457280159, 0.21181727945804596, -0.07436583936214447, -0.06398320943117142, -0.12686648964881897, 0.1733897477388382, -0.0044795009307563305, 0.026220498606562614, 0.011539910919964314, -0.09374070912599564, 0.030329275876283646, 0.19051384925842285, 0.1631309986114502, -0.08902175724506378, 0.006184775847941637, -0.03405250236392021, -0.010518989525735378, -0.062219105660915375, 0.07577589154243469, 0.11887829005718231, -0.005887000821530819, -0.06561030447483063, -0.029204271733760834, -0.0681641548871994, -0.028673192486166954, 0.0056154136545956135, 0.07179362326860428, 0.018411917611956596, -0.0031227271538227797, -0.03962535038590431, 0.09613676369190216, -0.06375392526388168, -0.1339055448770523, 0.01962723210453987, -0.1709652692079544, -0.14274592697620392, -0.03785032406449318, 0.09807087481021881, 0.03430141136050224, 0.03357571363449097, -0.03962371125817299, -0.004926025867462158, 0.08929938077926636, -0.013342314399778843, -0.042128536850214005, -0.09605023264884949, 0.054669443517923355, -0.1013873815536499, 0.2195015847682953, -0.03889377787709236, 0.047959376126527786, 0.10732451826334, 0.047696277499198914, -0.08411876857280731, 0.10550010204315186, 0.04659489169716835, -0.08619684725999832, -0.0001399270840920508, 0.12879131734371185, -0.05598938465118408, 0.16544625163078308, 0.050010696053504944, -0.15535761415958405, -0.004974417854100466, -0.02059159427881241, -0.093654103577137, -0.04530417174100876, -0.05095282196998596, -0.06085655465722084, 0.1248985305428505, 0.13296858966350555, -0.05335758253931999, -0.00032151155755855143, -0.047267068177461624, 0.053566668182611465, 0.08782485872507095, 0.028427185490727425, -0.012958439067006111, -0.2530754506587982, 0.036711812019348145, 0.05588744580745697, -0.008578677661716938, -0.26488491892814636, -0.08227638900279999, -0.02446601539850235, -0.04681609943509102, -0.0766659751534462, 0.06966593861579895, 0.10631562024354935, 0.042047057300806046, -0.06713500618934631, -0.0601264052093029, -0.07116956263780594, 0.17082804441452026, -0.11559001356363297, -0.10621830821037292 ]
null
null
transformers
this is [miqu-1-70b](https://huggingface.co/miqudev/miqu-1-70b), dequantised from q5 to f16 && transposed to pytorch. shapes have been rotated less wrongly than in [alpindale/miqu-1-70b-pytorch](https://huggingface.co/alpindale/miqu-1-70b-pytorch/tree/main) usage ```python from transformers import LlamaForCausalLM as LLM, LlamaTokenizer as LT lt = LT.from_pretrained("NousResearch/Llama-2-7b-hf") t = lt("[INST] eloquent high camp prose about a cute catgirl [/INST]", return_tensors='pt').input_ids.cuda() llm = LLM.from_pretrained("152334H/miqu-1-70b-sf", device_map='auto') # note: you may need many gpus for this out = llm.generate(t, use_cache=False, max_new_tokens=200) print(lt.decode(out[0])) ``` result: ``` <s> [INST] eloquent high camp prose about a cute catgirl [/INST] In the resplendent realm of high camp, where irony and extravagance dance in a dazzling pas de deux, there exists a creature of such enchanting allure that she captivates the hearts and minds of all who behold her. This beguiling figure, a vision of feline grace and innocence, is none other than the inimitable catgirl. With her delicate features and winsome smile, she is the embodiment of a dream, a living testament to the power of imagination and the boundless possibilities of the human spirit. Her eyes, those twin orbs of sapphire fire, sparkle with a mischievous intelligence that belies her diminutive stature. They are windows into a soul that is at once ancient and eternally young, a soul that has traversed the vast expanse of time and space to find solace in the warm embrace of human companion ``` this roughly (but not entirely) matches the llama.cpp q5 result: ```bash $ ./main -ngl 99 -m ./miqu-*q5* --color --temp 0.0 -n -1 -p '[INST] eloquent high camp prose about a cute catgirl [/INST]' ... [INST] eloquent high camp prose about a cute catgirl [/INST] In the resplendent realm of high camp, where irony and extravagance dance in a dazzling pas de deux, there exists a creature so enchantingly adorable that she captures the hearts of all who behold her. This is no ordinary feline, but rather a vision of elegance and whimsy combined: the cute catgirl. With her delicate features framed by an ethereal halo of pastel tresses, this darling diva prowls through life with the grace of a prima ballerina and the playfulness of a kitten. Her eyes, twin pools of sapphire or emerald, sparkle with mischief and intelligence as they survey their surroundings, ever alert for the next grand adventure or delightful prank. Her ensemble is a symphony of ruffles, bows, and lace, each detail painstakingly chosen to accentuate her lithe form and play up her feline charms. A frilly apron adorned with paw prints sways gently as she moves, while dainty ears perched atop her head twitch in response to every sound. Her gloved hands, so petite and perfect, seem made for holding teacups or sketching delicate portraits of her many admirers. But do not be fooled by her diminutive stature and sweet demeanor; beneath that fluffy exterior lies the heart of a lioness. Fiercely loyal and protective, she will stop at nothing to defend those she loves from harm. And when the situation calls for it, she can unleash a ferocious roar that belies her cute exterior. Indeed, the cute catgirl is a paradox wrapped in ruffles and ribbons, a living embodiment of the high camp aesthetic. She revels in the absurdity of her existence, finding joy in every outrageous situation and turning even the most mundane tasks into opportunities for sartorial expression. In her world, there is no such thing as too much glitter or too many bows; more is always more, and excess is a virtue to be celebrated. So let us raise our teacups in honor of this fabulous feline, this queen of camp who reminds us that life is too short for dull clothing and boring hairstyles. May we all strive to embody her spirit, embracing the absurdity of existence with open arms and a generous helping of glitter. Long live the cute catgirl! [end of text] ``` exl2 3.0bpw coming soon ![](https://thicc-af.mywaifulist.moe/waifus/miku-nakano-the-quintessential-quintuplets/phUEiEhPOL75GTDLncGy2dUbkDVMfYExZ2A1RBeQ.png?class=thumbnail)
{"license": "mit"}
text-generation
LoneStriker/miqu-1-70b-sf-GPTQ
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "license:mit", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T11:51:10+00:00
[]
[]
TAGS #transformers #safetensors #llama #text-generation #conversational #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
this is miqu-1-70b, dequantised from q5 to f16 && transposed to pytorch. shapes have been rotated less wrongly than in alpindale/miqu-1-70b-pytorch usage result: this roughly (but not entirely) matches the URL q5 result: exl2 3.0bpw coming soon ![](URL
[]
[ "TAGS\n#transformers #safetensors #llama #text-generation #conversational #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 56 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #conversational #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.0005691677215509117, 0.03026840090751648, -0.005779425613582134, -0.007680800743401051, 0.12393421679735184, -0.016670791432261467, 0.20057982206344604, 0.10256171226501465, 0.0027368133887648582, -0.025139721110463142, 0.14202554523944855, 0.2090049833059311, -0.02653181366622448, 0.03626692295074463, -0.12612687051296234, -0.17803461849689484, 0.08635426312685013, -0.0035030082799494267, 0.04833058640360832, 0.08384886384010315, 0.10781697928905487, -0.06203937530517578, 0.08011354506015778, -0.04676711559295654, -0.0940304547548294, 0.03246979042887688, 0.06822408735752106, -0.13098374009132385, 0.12263950705528259, 0.06911366432905197, 0.10126137733459473, 0.060711462050676346, -0.04248986020684242, -0.24028536677360535, 0.02768733911216259, -0.005825731437653303, -0.06894944608211517, 0.02077638730406761, 0.03447159752249718, -0.06464606523513794, 0.06811448931694031, 0.059148240834474564, -0.0005908439634367824, 0.09489285200834274, -0.1416224092245102, 0.002021060325205326, -0.04540175572037697, -0.031125035136938095, 0.11837171018123627, 0.08212298899888992, -0.005940968636423349, 0.11567636579275131, -0.05774720758199692, 0.09672804176807404, 0.0878739207983017, -0.3562628924846649, 0.01722630485892296, 0.13177140057086945, 0.08636787533760071, 0.04357139766216278, -0.06664732843637466, 0.10845058411359787, 0.07295350730419159, -0.03431075066328049, 0.03404705971479416, -0.0736931562423706, -0.060907237231731415, 0.05151093006134033, -0.05455131456255913, -0.024238532409071922, 0.23274658620357513, -0.0344974659383297, 0.027171114459633827, -0.0938771441578865, -0.07156997174024582, -0.0049468339420855045, -0.026434529572725296, 0.018185000866651535, -0.013636891730129719, 0.09878148138523102, 0.022968489676713943, -0.03718412294983864, -0.13966959714889526, -0.015280310995876789, -0.1799161285161972, 0.13825315237045288, 0.008047109469771385, 0.0309466402977705, -0.17635370790958405, 0.03744986280798912, 0.03522920608520508, -0.09773970395326614, -0.011200277134776115, -0.06822812557220459, 0.055764082819223404, -0.010668330825865269, -0.03837919235229492, -0.07488340884447098, 0.1452808827161789, 0.1400337964296341, -0.004895935300737619, 0.013622444123029709, -0.10354432463645935, 0.087824247777462, 0.005094456952065229, 0.02972278743982315, 0.059570424258708954, -0.03287388011813164, 0.06610265374183655, -0.10489657521247864, 0.06430593878030777, -0.04465807229280472, -0.15752150118350983, 0.024476690217852592, 0.0024344215635210276, 0.14078456163406372, 0.0003116365405730903, 0.10022792965173721, -0.0461345873773098, 0.05870504677295685, 0.02821454033255577, -0.07917805761098862, -0.007815105840563774, 0.0045831771567463875, 0.05050428956747055, 0.03022189997136593, 0.009842618368566036, 0.04734061658382416, -0.046343643218278885, 0.040167130529880524, -0.0721283107995987, -0.04361659660935402, -0.047072943300008774, -0.05691814795136452, 0.054554544389247894, -0.04412351921200752, 0.031097736209630966, -0.17774003744125366, -0.20105750858783722, 0.006062032654881477, 0.014256932772696018, -0.02251455932855606, -0.01111325342208147, -0.05564434826374054, -0.03297819197177887, 0.030923429876565933, -0.08253207057714462, -0.08266308903694153, -0.08656799048185349, 0.08184840530157089, -0.034670811146497726, 0.04964166879653931, -0.13810113072395325, 0.04346073791384697, -0.1097627505660057, 0.024651816114783287, -0.02479342184960842, 0.04081350192427635, -0.037637315690517426, 0.17413905262947083, -0.014600466005504131, 0.036081500351428986, -0.06184709072113037, 0.08253529667854309, -0.047251146286726, 0.20924541354179382, -0.13303455710411072, -0.045243039727211, 0.21816439926624298, -0.11933718621730804, -0.2199726104736328, 0.09134338796138763, -0.017768139019608498, 0.10163438320159912, 0.13150125741958618, 0.2125980705022812, -0.003975315950810909, -0.08655516803264618, 0.05120107904076576, 0.09330737590789795, -0.07744380831718445, -0.10532614588737488, 0.006278107408434153, -0.0237810667604208, -0.1067701205611229, 0.030834544450044632, 0.09159471839666367, 0.055198732763528824, -0.0206023920327425, -0.060277264565229416, -0.044726673513650894, -0.03647027537226677, -0.006402628030627966, -0.034374698996543884, 0.05666901171207428, -0.10561396181583405, 0.0008384078973904252, 0.04050183296203613, -0.009274031966924667, -0.010979763232171535, 0.018539749085903168, -0.09420713037252426, 0.06333526968955994, 0.010708075948059559, 0.0486791618168354, -0.09103409945964813, -0.1205744668841362, -0.026474347338080406, 0.1048177108168602, 0.014087734743952751, 0.057943567633628845, 0.03745978698134422, -0.012622485868632793, -0.013452895916998386, 0.010302813723683357, 0.21106989681720734, 0.03879684954881668, -0.055580686777830124, -0.09376689791679382, 0.1019076481461525, -0.06258829683065414, 0.014857032336294651, -0.10959229618310928, 0.024967070668935776, 0.05571519583463669, 0.09035582095384598, 0.018503490835428238, 0.061780765652656555, -0.0193456020206213, 0.0021762927062809467, -0.08686630427837372, 0.006882676389068365, 0.09584054350852966, 0.010849484242498875, -0.1046229675412178, 0.22536177933216095, -0.2302411049604416, 0.25318604707717896, 0.1955074816942215, -0.18912889063358307, 0.022286945953965187, -0.11619117856025696, -0.0030349870212376118, 0.01123755518347025, 0.019197653979063034, -0.052956365048885345, 0.020005550235509872, -0.006996720563620329, 0.17232349514961243, -0.06575432419776917, -0.03044302575290203, -0.02386540174484253, -0.07438083738088608, -0.051245078444480896, 0.060514211654663086, 0.09873983263969421, -0.1872550994157791, 0.18996097147464752, 0.23916521668434143, 0.04921855032444, 0.18687786161899567, -0.04146864637732506, 0.015993451699614525, 0.03823651000857353, 0.04348839819431305, -0.00034227073774673045, -0.02700083889067173, -0.12750719487667084, -0.014179466292262077, 0.05456198379397392, 0.021291760727763176, 0.07244087010622025, -0.12852837145328522, -0.07239031791687012, -0.0019381374586373568, -0.0347394123673439, 0.0321790836751461, 0.07742355763912201, 0.002989932894706726, 0.12464089691638947, -0.05088387429714203, -0.06383717805147171, 0.11249808222055435, -0.02096649445593357, -0.09606146812438965, 0.18102820217609406, -0.13335320353507996, -0.2577352821826935, -0.16749133169651031, -0.1758090853691101, -0.04561625048518181, 0.05666625499725342, 0.12794378399848938, -0.05771476402878761, -0.06537273526191711, -0.06109621748328209, 0.028574729338288307, -0.020225634798407555, -0.005206136032938957, -0.04940271005034447, 0.06738027930259705, -0.06819579005241394, -0.11573685705661774, -0.05808679759502411, 0.017959145829081535, -0.06612885743379593, 0.10544807463884354, -0.08685622364282608, 0.08593754470348358, 0.14343073964118958, 0.023149555549025536, 0.0056336382403969765, -0.06615734100341797, 0.14803656935691833, -0.06705351918935776, -0.004510159138590097, 0.18866874277591705, -0.04495077580213547, 0.056410688906908035, 0.19779126346111298, 0.019139746204018593, -0.1258988231420517, 0.058476053178310394, -0.01870557852089405, -0.09270547330379486, -0.24951080977916718, -0.1107277199625969, -0.09696482121944427, 0.09810655564069748, 0.004071237985044718, 0.07468543201684952, 0.16529400646686554, 0.0666000097990036, -0.029285557568073273, -0.03729884698987007, 0.07551243156194687, 0.09431212395429611, 0.28239867091178894, -0.04563182219862938, 0.13355515897274017, -0.09407903999090195, -0.12141021341085434, 0.06808027625083923, 0.08666377514600754, 0.09740467369556427, 0.14567551016807556, 0.10435792803764343, 0.061198752373456955, 0.052402377128601074, 0.11356253176927567, 0.10255523771047592, 0.03482705354690552, -0.045480839908123016, -0.02383856289088726, -0.04675186425447464, -0.023478806018829346, 0.06876002997159958, -0.057931844145059586, -0.1251164674758911, -0.02431958168745041, -0.05684112757444382, 0.08322072774171829, 0.10292354971170425, 0.04484309256076813, -0.19743771851062775, 0.027478473260998726, 0.1391770988702774, -0.033375564962625504, -0.08757176250219345, 0.12839633226394653, 0.04147679731249809, -0.05758469179272652, 0.08718124777078629, -0.02552475407719612, 0.10689202696084976, -0.022716548293828964, 0.0845961943268776, -0.09373506158590317, -0.09592817723751068, -0.002121838042512536, 0.10605843365192413, -0.33184942603111267, 0.21520109474658966, 0.017110098153352737, -0.0011029692832380533, -0.06920389831066132, -0.007827462628483772, 0.007449707947671413, 0.1494879573583603, 0.15354286134243011, -0.036155834794044495, -0.12665162980556488, -0.04870329052209854, -0.00755130173638463, 0.027197029441595078, 0.11079981178045273, 0.0005214710836298764, 0.008046012371778488, -0.07294304668903351, 0.0031203555408865213, 0.011225464753806591, -0.03095840848982334, -0.03996356576681137, -0.18460363149642944, 0.0215305108577013, 0.1286952793598175, 0.10708256810903549, -0.027399228885769844, 0.02612617425620556, -0.13613013923168182, 0.1831846833229065, -0.1099759116768837, -0.052929844707250595, -0.11382761597633362, -0.14973537623882294, 0.000214276704355143, -0.02133144997060299, 0.052957307547330856, -0.06085783615708351, 0.04300437122583389, -0.0991264060139656, -0.18037806451320648, 0.1218428760766983, -0.1038995236158371, -0.02295663207769394, -0.052425630390644073, 0.16215817630290985, -0.09474063664674759, -0.022231630980968475, 0.06611523777246475, 0.035305205732584, -0.05566869303584099, -0.09669981896877289, -0.009570736438035965, 0.0026432990562170744, 0.029019128531217575, -0.015986355021595955, -0.1340111792087555, -0.08610024303197861, -0.011368979699909687, -0.07684875279664993, 0.2708785831928253, 0.25187867879867554, -0.04344089329242706, 0.16429945826530457, 0.1766400933265686, -0.11247728765010834, -0.35223308205604553, -0.1094401404261589, -0.20467765629291534, -0.05295013636350632, 0.017857976257801056, -0.102334164083004, 0.08513185381889343, 0.029009856283664703, -0.04985405132174492, 0.10192443430423737, -0.2161361426115036, -0.116386279463768, 0.16906258463859558, 0.032223232090473175, 0.3679789900779724, -0.19134950637817383, -0.11310295760631561, -0.11911448836326599, -0.09437704086303711, 0.16245406866073608, -0.1050034686923027, 0.0903928354382515, 0.026932088658213615, 0.06107216328382492, 0.03261984884738922, -0.03233310952782631, 0.11524650454521179, -0.03482987731695175, 0.06873463839292526, -0.13058903813362122, -0.016216721385717392, 0.0364764928817749, -0.020026322454214096, 0.052481792867183685, -0.16696332395076752, 0.01610221154987812, -0.04988480359315872, -0.04296071082353592, -0.011710057035088539, 0.08053705841302872, 0.006171372253447771, -0.06921958923339844, -0.03883899003267288, -0.0534801259636879, 0.01946091279387474, -0.01371228788048029, 0.2759011387825012, -0.08533263951539993, 0.1845393031835556, 0.1678970456123352, 0.16199782490730286, -0.12405289709568024, 0.09155070036649704, -0.03589274734258652, -0.08797135949134827, 0.07076383382081985, -0.12157030403614044, 0.05676010996103287, 0.08774981647729874, -0.047847531735897064, 0.102900929749012, 0.07471447438001633, 0.020537683740258217, -0.005388888996094465, 0.14900201559066772, -0.1911458671092987, -0.10747314989566803, -0.029038699343800545, 0.08607541024684906, 0.08602768927812576, 0.08265142887830734, 0.18157722055912018, -0.011384707875549793, 0.02875511534512043, 0.0002349212736589834, 0.04147034510970116, -0.04067661240696907, 0.034632571041584015, 0.004870852455496788, 0.019318271428346634, -0.12345778942108154, 0.1052570641040802, 0.004227067343890667, -0.12422677874565125, 0.030803296715021133, 0.09900417923927307, -0.11316242069005966, -0.12551146745681763, -0.061151985079050064, 0.14257317781448364, -0.14595121145248413, -0.07654847949743271, -0.043176110833883286, -0.1870022714138031, 0.030991066247224808, 0.22905074059963226, 0.03893444687128067, 0.10527749359607697, 0.00025028636446222663, -0.047077130526304245, -0.025180570781230927, 0.04380583390593529, -0.07184609770774841, 0.020637260749936104, -0.10354288667440414, -0.012115558609366417, -0.035101260989904404, 0.046457234770059586, -0.08753124624490738, -0.03981354087591171, -0.16556434333324432, 0.034135427325963974, -0.15149405598640442, -0.02009320817887783, -0.10041918605566025, -0.026495076715946198, 0.021843044087290764, -0.018219182267785072, -0.05821055546402931, -0.05869567394256592, -0.09869731217622757, 0.02559635415673256, -0.020244602113962173, 0.07858848571777344, -0.10507214069366455, -0.029194623231887817, 0.06234395131468773, -0.02745119296014309, 0.10503295063972473, 0.050207268446683884, -0.09599626064300537, 0.08977745473384857, -0.2548978626728058, -0.03408842533826828, 0.12418515235185623, -0.0024804214481264353, 0.021922416985034943, 0.05795785412192345, -0.01517826784402132, 0.1253129094839096, 0.016869723796844482, 0.056829195469617844, -0.03426999971270561, -0.11613040417432785, 0.017578652128577232, -0.02382705546915531, -0.12108992040157318, -0.015287832356989384, -0.08777373284101486, 0.07894490659236908, -0.045772697776556015, 0.17387144267559052, -0.079983189702034, 0.044346775859594345, -0.03832852840423584, 0.03650568798184395, 0.016848308965563774, -0.17066964507102966, -0.10868518799543381, -0.08728142827749252, -0.011061674915254116, -0.008855300024151802, 0.28512516617774963, 0.02954568900167942, -0.07765054702758789, 0.08719948679208755, 0.04720422998070717, 0.03436547890305519, 0.03705444186925888, 0.2878183424472809, 0.09032974392175674, -0.023173412308096886, -0.14650490880012512, 0.02621215023100376, 0.020107239484786987, -0.0833572968840599, 0.07981549948453903, 0.10982800275087357, -0.08972351998090744, 0.10125767439603806, 0.05650821700692177, 0.00820234790444374, -0.013232419267296791, -0.08556391298770905, -0.036529455333948135, 0.04482261463999748, -0.040347617119550705, 0.08681420981884003, 0.2188776284456253, -0.04128800332546234, 0.012239835225045681, -0.05264842137694359, -0.03152349218726158, -0.19268813729286194, -0.13786496222019196, -0.09619536995887756, -0.11100079119205475, 0.027774536982178688, -0.08766230940818787, 0.05614083632826805, 0.06316492706537247, 0.06389103829860687, -0.042949922382831573, 0.09820075333118439, -0.02900860644876957, -0.05136289447546005, 0.02691546082496643, -0.029937779530882835, 0.05826915055513382, -0.05261525139212608, -0.04776594415307045, -0.07043541222810745, -0.037922609597444534, -0.04753337800502777, 0.07690087705850601, 0.020815400406718254, 0.05223739147186279, -0.15359240770339966, -0.06513933837413788, -0.03139248117804527, 0.07923950999975204, -0.03322955593466759, 0.1384638249874115, 0.015548406168818474, -0.032195813953876495, 0.07704564929008484, 0.2116333395242691, -0.05921745300292969, -0.10795019567012787, -0.02262108400464058, 0.17396169900894165, 0.02381749264895916, 0.13749580085277557, -0.051188550889492035, -0.007439639884978533, -0.027830881997942924, 0.32955774664878845, 0.2775338590145111, -0.06558609753847122, 0.027906622737646103, -0.06983242183923721, 0.043349675834178925, 0.06356148421764374, 0.1187676340341568, 0.06889500468969345, 0.2647046148777008, -0.038245122879743576, -0.009535894729197025, -0.00723807979375124, -0.011412912979722023, -0.12783683836460114, 0.1092628464102745, -0.008579141460359097, -0.04360508173704147, -0.024161534383893013, 0.11072532087564468, -0.18454989790916443, 0.07873725146055222, -0.06403639167547226, -0.11388422548770905, -0.006975886411964893, -0.007217585574835539, 0.1705339103937149, -0.020880691707134247, 0.041406676173210144, -0.02618410252034664, -0.08844110369682312, 0.010316282510757446, 0.0013190775644034147, -0.20132848620414734, 0.007113540079444647, 0.0317230261862278, 0.0026388487312942743, 0.06414279341697693, -0.005612512584775686, 0.03002963773906231, 0.07572976499795914, 0.0228758342564106, -0.04426576569676399, 0.15260286629199982, 0.036797065287828445, -0.08339497447013855, 0.045644111931324005, -0.0406322218477726, -0.0022905210498720407, 0.01915745809674263, 0.060521602630615234, -0.11291439086198807, 0.06985818594694138, -0.002728865947574377, -0.11687634885311127, -0.03295402601361275, 0.03366249427199364, -0.07441430538892746, 0.07509835809469223, 0.020866960287094116, -0.017463620752096176, 0.011251081712543964, -0.024018414318561554, 0.011885357089340687, -0.006544362287968397, -0.16443650424480438, -0.02983039617538452, -0.12943366169929504, -0.0619380846619606, 0.12867197394371033, 0.011144373565912247, -0.2688652575016022, 0.003814426949247718, -0.09708355367183685, 0.06670353561639786, -0.18629629909992218, 0.059078771620988846, 0.20747289061546326, 0.01242530345916748, -0.022893745452165604, -0.17272116243839264, 0.06423069536685944, 0.07249844819307327, -0.037798795849084854, -0.09726443141698837 ]
null
null
null
# **Q-Learning** Agent playing1 **FrozenLake-v1** This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** . ## Usage ```python model = load_from_hub(repo_id="tahsinkheya/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
{"tags": ["FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "q-FrozenLake-v1-4x4-noSlippery", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "FrozenLake-v1-4x4-no_slippery", "type": "FrozenLake-v1-4x4-no_slippery"}, "metrics": [{"type": "mean_reward", "value": "1.00 +/- 0.00", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
tahsinkheya/q-FrozenLake-v1-4x4-noSlippery
[ "FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
2024-02-10T11:51:11+00:00
[]
[]
TAGS #FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
# Q-Learning Agent playing1 FrozenLake-v1 This is a trained model of a Q-Learning agent playing FrozenLake-v1 . ## Usage
[ "# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage" ]
[ "TAGS\n#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n", "# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage" ]
[ 40, 39 ]
[ "passage: TAGS\n#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage" ]
[ 0.04578453302383423, -0.08074592798948288, -0.00430759321898222, 0.10720831900835037, 0.05034215748310089, -0.040469273924827576, 0.11997015029191971, 0.018999949097633362, 0.20601962506771088, -0.010012076236307621, 0.1455274522304535, 0.007022971753031015, -0.006192410364747047, 0.1867983490228653, 0.04572829231619835, -0.26324528455734253, 0.01831899583339691, -0.09495259821414948, -0.07281816750764847, 0.11870454251766205, 0.05470194295048714, -0.01901467889547348, -0.0007633853238075972, 0.056141503155231476, -0.0673527717590332, 0.0007737681735306978, 0.031996939331293106, -0.012976245954632759, 0.19804789125919342, -0.02254498563706875, 0.06641989201307297, 0.054705578833818436, 0.0758768692612648, -0.1998077929019928, 0.0358855277299881, -0.04215473681688309, -0.09439758956432343, -0.03934839740395546, -0.018780618906021118, 0.05878105387091637, 0.053356342017650604, 0.03858819976449013, 0.058354366570711136, 0.09384993463754654, -0.0773480236530304, 0.04328357055783272, 0.04280758649110794, 0.024811049923300743, 0.04589218273758888, -0.0237203948199749, -0.027002155780792236, 0.08246652781963348, -0.22182892262935638, 0.10318073630332947, -0.010159241035580635, -0.5270710587501526, -0.00633762264624238, 0.24088262021541595, 0.11517096310853958, 0.05707438662648201, -0.06903956830501556, 0.10566288232803345, 0.03913382440805435, -0.007209456991404295, 0.03210983797907829, 0.02150118350982666, 0.12817370891571045, 0.06009242683649063, -0.09581366181373596, 0.040699947625398636, 0.13722525537014008, 0.012822695076465607, 0.020306183025240898, -0.08888901025056839, 0.0410032719373703, -0.03461858257651329, -0.007679527159780264, -0.09758518636226654, 0.05478060990571976, 0.012466507963836193, -0.0934976264834404, -0.09247440844774246, -0.04236573353409767, -0.06708304584026337, 0.11252415925264359, 0.046419668942689896, -0.0874939113855362, 0.03884070739150047, -0.06760413944721222, 0.05918780341744423, -0.16863860189914703, 0.02074250765144825, -0.06627868115901947, -0.09376336634159088, -0.11799788475036621, -0.01683047041296959, -0.07946427166461945, 0.009092256426811218, 0.056664444506168365, 0.1447116881608963, 0.22076484560966492, 0.06690320372581482, 0.09728849679231644, 0.07456006109714508, 0.06531001627445221, 0.1538129299879074, 0.10918238013982773, 0.019075315445661545, -0.015266558155417442, 0.0948706716299057, -0.06445580720901489, -0.1351388692855835, -0.15579092502593994, 0.005488025024533272, 0.0983937531709671, 0.08871900290250778, -0.044080477207899094, -0.006702381651848555, -0.024641724303364754, 0.08566431701183319, -0.11314457654953003, -0.024612564593553543, -0.002267979085445404, 0.06882024556398392, -0.024801667779684067, 0.020378148183226585, -0.06242705136537552, 0.12715265154838562, 0.04222423583269119, -0.059924717992544174, -0.055308472365140915, -0.03053177334368229, -0.014276440255343914, -0.027539284899830818, 0.02446848154067993, -0.07659092545509338, 0.04767750948667526, -0.16766095161437988, -0.042871296405792236, -0.04784649610519409, 0.025697942823171616, -0.03907240927219391, -0.13557587563991547, -0.17699143290519714, -0.048906855285167694, -0.022438718006014824, 0.03549358621239662, -0.038111843168735504, 0.006551501806825399, -0.006318534724414349, -0.1583600640296936, 0.09783563017845154, 0.09784027189016342, -0.03643378987908363, -0.02749447710812092, 0.056263517588377, -0.07194498926401138, 0.1561182290315628, -0.21054518222808838, -0.054014235734939575, -0.044764336198568344, -0.06595750898122787, 0.19673264026641846, 0.012690845876932144, -0.01202624011784792, 0.19873127341270447, -0.29073721170425415, -0.06078760325908661, 0.12533614039421082, -0.07834373414516449, -0.0936407670378685, 0.06941844522953033, -0.04206686094403267, 0.023345354944467545, 0.046047765761613846, 0.36345911026000977, -0.02069227211177349, -0.16197136044502258, -0.021782705560326576, 0.13971707224845886, -0.1184760183095932, 0.059895481914281845, 0.04240793362259865, 0.12543781101703644, -0.04250509291887283, -0.018672896549105644, -0.09023164212703705, 0.05999075248837471, -0.05241934582591057, -0.09016361832618713, -0.03393383324146271, -0.07645075023174286, 0.13294468820095062, -0.0629684180021286, 0.05601520463824272, -0.03255095332860947, -0.07133250683546066, -0.050324998795986176, -0.016492370516061783, 0.04460815340280533, 0.05951254442334175, -0.12794871628284454, 0.11029167473316193, 0.13025271892547607, -0.0006193425506353378, -0.07498852163553238, -0.17872096598148346, 0.003240168560296297, 0.009576505981385708, 0.039837226271629333, 0.17141658067703247, 0.12209978699684143, 0.033295199275016785, 0.008770671673119068, -0.06389404833316803, -0.18276847898960114, 0.058129217475652695, -0.056212130934000015, -0.14230976998806, -0.052409034222364426, -0.0728459507226944, 0.017381802201271057, -0.0859743058681488, -0.017379917204380035, 0.021926190704107285, 0.006908397190272808, 0.02990424446761608, -0.026645656675100327, -0.049561817198991776, 0.021254703402519226, 0.06490101665258408, -0.0037617047782987356, 0.12023693323135376, 0.008277264423668385, -0.18308481574058533, 0.07930773496627808, 0.08478537946939468, 0.09196605533361435, 0.013250201940536499, 0.02685922384262085, -0.021522263064980507, -0.08061408251523972, -0.054420311003923416, 0.02957955375313759, 0.11417073011398315, 0.1317172348499298, 0.2361993044614792, 0.08753683418035507, 0.04697408527135849, -0.02164587564766407, -0.016415923833847046, 0.002810494042932987, -0.06318057328462601, -0.029935607686638832, 0.10614971816539764, 0.05865858122706413, -0.067733034491539, -0.04576427489519119, 0.09590928256511688, 0.02732124738395214, 0.21205885708332062, -0.03342745825648308, 0.01286078616976738, -0.10957037657499313, -0.06550975888967514, -0.031982194632291794, 0.09201868623495102, 0.09498392790555954, 0.009755023755133152, -0.022056059911847115, -0.04259001836180687, 0.0012916827108711004, -0.1334889680147171, -0.10375088453292847, 0.026475343853235245, 0.013400445692241192, -0.11206940561532974, 0.11674030870199203, -0.11352457851171494, 0.039504457265138626, 0.06024791672825813, -0.13837239146232605, 0.04428480193018913, -0.029713207855820656, -0.07886212319135666, 0.16866780817508698, -0.11075661331415176, -0.094340018928051, -0.08831550180912018, 0.004082420375198126, 0.0075836325995624065, -0.03922267258167267, -0.009283260442316532, -0.19952571392059326, -0.005375816952437162, -0.03544965013861656, 0.013616434298455715, -0.06988783925771713, -0.11287739872932434, -0.010957922786474228, 0.07084179669618607, -0.043388739228248596, -0.07803605496883392, 0.007967432029545307, -0.08923084288835526, -0.10623309016227722, 0.028189711272716522, 0.019765101373195648, -0.022883659228682518, 0.16152891516685486, 0.01816628873348236, 0.05626589432358742, -0.03298520669341087, 0.30665266513824463, -0.038163769990205765, 0.08371731638908386, -0.02993497997522354, -0.07433546334505081, 0.06130730360746384, -0.022327827289700508, 0.06086638569831848, -0.020221687853336334, -0.02362890914082527, 0.0077952733263373375, -0.08579335361719131, -0.18365982174873352, -0.05417544022202492, 0.03724347800016403, 0.195254847407341, 0.031118987128138542, 0.01910330168902874, -0.0488768145442009, -0.010547760874032974, 0.1665220558643341, -0.10005921125411987, 0.04030545800924301, -0.05366240441799164, 0.11506262421607971, -0.08640182018280029, 0.06195629760622978, 0.020486772060394287, 0.04266135022044182, -0.04877188801765442, 0.09486009180545807, 0.0826394334435463, 0.1121082529425621, -0.02206910029053688, 0.046257395297288895, 0.019012698903679848, 0.07383184134960175, 0.11073657125234604, 0.0368414968252182, -0.0729052945971489, 0.001982470043003559, -0.006313489284366369, -0.039427030831575394, 0.11933320760726929, 0.17963355779647827, -0.11991413682699203, -0.05106910318136215, 0.27167606353759766, 0.0031242913100868464, 0.19481229782104492, -0.01315275114029646, 0.043591804802417755, -0.04484925419092178, 0.04572054371237755, -0.05338600277900696, -0.04086209088563919, 0.2094656229019165, 0.08045925945043564, -0.17165091633796692, -0.08549032360315323, -0.05912299454212189, 0.07081323862075806, 0.10728751868009567, 0.0013539529172703624, -0.04156802222132683, 0.0004610282776411623, 0.0014198932331055403, 0.08339415490627289, -0.14520122110843658, 0.11816094070672989, -0.03172019124031067, 0.05612684786319733, 0.017555562779307365, -0.045326150953769684, 0.04264266416430473, 0.07474290579557419, 0.26618310809135437, 0.0904107540845871, -0.040318213403224945, -0.0892091691493988, -0.12260187417268753, 0.010461576282978058, 0.029102616012096405, -0.03534553572535515, 0.0037547778338193893, -0.020087555050849915, 0.0318896509706974, 0.008264793083071709, 0.016230624169111252, -0.08987458795309067, -0.03175399824976921, -0.027736429125070572, -0.023839212954044342, 0.10733365267515182, -0.09495144337415695, -0.1444292515516281, -0.15713949501514435, 0.04191131144762039, -0.0766405463218689, -0.056593164801597595, -0.054507751017808914, -0.05239389091730118, -0.0311186034232378, -0.03773957118391991, 0.09099467098712921, -0.0021037792321294546, 0.14807306230068207, -0.1920108050107956, -0.04220759496092796, 0.051812779158353806, -0.07607918977737427, -0.08729588985443115, 0.03410962224006653, 0.12136995792388916, 0.05116051807999611, 0.11504370719194412, 0.013609255664050579, 0.09567681699991226, 0.0045484392903745174, -0.06713183224201202, 0.15302421152591705, -0.14069625735282898, -0.27875974774360657, -0.03836318850517273, 0.016946332529187202, 0.1615200787782669, -0.05613167956471443, 0.031766023486852646, 0.3335736393928528, 0.27782970666885376, -0.1428707242012024, 0.25916144251823425, 0.019178593531250954, 0.004398873541504145, -0.19130495190620422, -0.10125631093978882, 0.025324683636426926, 0.04740457236766815, 0.12032642960548401, -0.14564448595046997, -0.010732659138739109, -0.04543145373463631, -0.025908485054969788, 0.10386138409376144, -0.12300799041986465, -0.07263197749853134, 0.07765276730060577, 0.039809420704841614, 0.1808302253484726, 0.03932500258088112, 0.0014799144119024277, 0.13626977801322937, 0.06612244248390198, 0.019124457612633705, 0.05216038227081299, 0.08028066903352737, -0.018944554030895233, 0.14207926392555237, 0.05448179319500923, -0.02551644667983055, 0.052681710571050644, -0.0054580713622272015, -0.03219012916088104, 0.015605825930833817, -0.183198019862175, -0.10147556662559509, -0.0561356320977211, -0.10798973590135574, -0.04978342354297638, 0.056853994727134705, -0.12395523488521576, -0.007896827533841133, -0.03841273859143257, 0.03718273714184761, -0.07831971347332001, -0.09360362589359283, -0.036494381725788116, 0.1351792961359024, 0.07210618257522583, 0.04471297934651375, 0.035655103623867035, -0.07390819489955902, 0.07097936421632767, 0.21671734750270844, 0.08159157633781433, 0.028919655829668045, -0.19545674324035645, -0.024042490869760513, -0.0803457647562027, 0.06306298077106476, -0.08856996893882751, -0.016788700595498085, 0.11923003196716309, 0.08616556972265244, 0.05413002520799637, 0.09640096127986908, -0.045083072036504745, 0.021686913445591927, 0.02684609219431877, -0.15131035447120667, -0.18501274287700653, -0.08534606546163559, -0.03519878163933754, 0.11561143398284912, -0.06398691236972809, 0.10897188633680344, -0.13615410029888153, 0.010051886551082134, -0.006060056854039431, 0.02693452313542366, -0.03596206381917, -0.11251141875982285, 0.15348562598228455, 0.11999429017305374, -0.06767056882381439, 0.03127254918217659, -0.09527092427015305, -0.04423454403877258, 0.12686803936958313, -0.013623855076730251, -0.0371493324637413, -0.054547641426324844, -0.03628576174378395, 0.15247689187526703, -0.03436964750289917, 0.008244883269071579, -0.041229065507650375, -0.18217355012893677, 0.0798322781920433, 0.09045056998729706, 0.019827889278531075, -0.031874191015958786, -0.09797266125679016, -0.010231015272438526, -0.0011165260802954435, 0.11730700731277466, -0.10696814209222794, -0.10933240503072739, -0.15144047141075134, 0.06713984161615372, -0.0007159380475059152, 0.18502596020698547, -0.06394898891448975, -0.08904669433832169, -0.12429379671812057, 0.02344517596065998, -0.0027384376153349876, -0.042264558374881744, 0.01618490368127823, 0.07992301136255264, -0.04095321521162987, 0.02075677551329136, -0.06651144474744797, 0.06372585147619247, -0.11786920577287674, 0.09625071287155151, 0.01063506118953228, 0.016993753612041473, -0.0417880080640316, -0.01618220843374729, 0.039470795542001724, -0.057925306260585785, 0.07921463251113892, 0.011758086271584034, 0.0010938759660348296, 0.10196787863969803, -0.0034960443153977394, 0.06409632414579391, -0.05372481048107147, -0.023290161043405533, 0.06578411161899567, -0.05874887853860855, -0.03370826691389084, -0.1573946475982666, -0.0709633082151413, 0.020051732659339905, -0.04775108024477959, 0.002077929675579071, 0.03673801198601723, 0.062159497290849686, -0.06937079131603241, -0.12125655263662338, -0.043812792748212814, -0.028638383373618126, 0.021301284432411194, 0.10829301923513412, -0.07526551932096481, 0.1547859013080597, -0.052787959575653076, -0.00020603960729204118, 0.07437096536159515, 0.04048224538564682, 0.01393822580575943, -0.10422444343566895, -0.04698587954044342, -0.11035211384296417, 0.1502903699874878, -0.007902312092483044, -0.03533121198415756, 0.03719403222203255, -0.11946307867765427, -0.1572723090648651, 0.03418220207095146, 0.10199101269245148, 0.0448341928422451, 0.025807438418269157, 0.027079269289970398, -0.04042419046163559, -0.021270349621772766, -0.07034418731927872, 0.0882953479886055, -0.12085357308387756, -0.09669415652751923, 0.09555385261774063, 0.12178351730108261, -0.0036850625183433294, -0.07441367954015732, 0.11554073542356491, -0.021787192672491074, 0.05525410920381546, -0.02971339225769043, 0.10308072715997696, 0.0796005055308342, -0.12273547053337097, 0.005693064536899328, -0.036891788244247437, -0.0741485133767128, -0.12975730001926422, 0.019545545801520348, -0.061916105449199677, -0.13383042812347412, 0.12179028987884521, -0.09376577287912369, 0.030037038028240204, -0.10506992787122726, 0.021338803693652153, 0.01864001713693142, 0.061665527522563934, -0.10988292098045349, 0.08575301617383957, 0.13424484431743622, -0.043199893087148666, -0.07184189558029175, -0.12455986440181732, -0.05022053420543671, -0.04231856390833855, -0.13957437872886658, -0.11600435525178909, 0.0100301094353199, -0.023418782278895378, -0.05818291753530502, 0.0015462689334526658, -0.03659068048000336, 0.008594646118581295, 0.021907730028033257, 0.04032021388411522, -0.02693161368370056, 0.05134565755724907, -0.057569269090890884, -0.052510857582092285, 0.11489357799291611, 0.04113486409187317, -0.03561042994260788, -0.052359987050294876, 0.12997733056545258, -0.11959461867809296, 0.07662346214056015, -0.020313527435064316, 0.017129231244325638, -0.06435854732990265, 0.17131924629211426, 0.11673715710639954, -0.1367570012807846, -0.005008010193705559, -0.08210669457912445, 0.020409544929862022, 0.023555370047688484, 0.13693512976169586, -0.03411718085408211, -0.0012358218664303422, -0.1580323874950409, 0.018575575202703476, -0.18557456135749817, -0.03716109320521355, 0.04671547934412956, 0.09917585551738739, 0.15293832123279572, -0.0034432117827236652, -0.1263325810432434, 0.10424192249774933, -0.2118520885705948, 0.0907607227563858, 0.05121984705328941, -0.11874113976955414, -0.06765396893024445, -0.06795281916856766, 0.1198519766330719, 0.009196433238685131, 0.2040700763463974, -0.013615905307233334, -0.09132910519838333, -0.07060808688402176, -0.01980910450220108, -0.030524181202054024, 0.09714830666780472, 0.041414931416511536, 0.04653804749250412, 0.12821412086486816, 0.00368314771912992, 0.07533777505159378, 0.060310911387205124, 0.02759413793683052, -0.012300663627684116, 0.04076618701219559, 0.08261215686798096, -0.14588621258735657, -0.1659701019525528, 0.1326720416545868, 0.025149408727884293, 0.11792458593845367, 0.03658788278698921, -0.1549617499113083, 0.06687124073505402, 0.2523096203804016, -0.11147607117891312, 0.02505038119852543, 0.12737524509429932, -0.0366884209215641, 0.0672016367316246, 0.1144871786236763, -0.02633814327418804, -0.05217865854501724, -0.011363590136170387, 0.10233135521411896, 0.028660254552960396, -0.04646271467208862, -0.02340836264193058, -0.03373933956027031, -0.019070526584982872, -0.011738128960132599, -0.0909019410610199, -0.1543993502855301, -0.10471053421497345, -0.16619662940502167, 0.04399140924215317, -0.04626438021659851, 0.13418889045715332, 0.09469578415155411, -0.012723101302981377, 0.04568437114357948, 0.028575526550412178, 0.07275456190109253, 0.07916246354579926, -0.02939477376639843, -0.036159269511699677 ]
null
null
peft
## Training procedure The following `bitsandbytes` quantization config was used during training: - load_in_8bit: False - load_in_4bit: True - llm_int8_threshold: 6.0 - llm_int8_skip_modules: None - llm_int8_enable_fp32_cpu_offload: False - llm_int8_has_fp16_weight: False - bnb_4bit_quant_type: nf4 - bnb_4bit_use_double_quant: False - bnb_4bit_compute_dtype: float16 ### Framework versions - PEFT 0.4.0
{"library_name": "peft"}
null
saracandu/llama2-7b-harrypotter
[ "peft", "region:us" ]
2024-02-10T11:52:24+00:00
[]
[]
TAGS #peft #region-us
## Training procedure The following 'bitsandbytes' quantization config was used during training: - load_in_8bit: False - load_in_4bit: True - llm_int8_threshold: 6.0 - llm_int8_skip_modules: None - llm_int8_enable_fp32_cpu_offload: False - llm_int8_has_fp16_weight: False - bnb_4bit_quant_type: nf4 - bnb_4bit_use_double_quant: False - bnb_4bit_compute_dtype: float16 ### Framework versions - PEFT 0.4.0
[ "## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: False\n- bnb_4bit_compute_dtype: float16", "### Framework versions\n\n\n- PEFT 0.4.0" ]
[ "TAGS\n#peft #region-us \n", "## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: False\n- bnb_4bit_compute_dtype: float16", "### Framework versions\n\n\n- PEFT 0.4.0" ]
[ 9, 154, 11 ]
[ "passage: TAGS\n#peft #region-us \n## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: False\n- bnb_4bit_compute_dtype: float16### Framework versions\n\n\n- PEFT 0.4.0" ]
[ -0.06979092955589294, 0.021877916529774666, -0.0024098597932606936, 0.1378258466720581, 0.10862318426370621, 0.07474296540021896, 0.10942144691944122, 0.12339992076158524, 0.0572756864130497, 0.09055759757757187, 0.09443594515323639, 0.05080708488821983, 0.07617583870887756, 0.15425844490528107, -0.02472618594765663, -0.030937515199184418, 0.05461447685956955, 0.0024960634764283895, 0.003683591727167368, 0.07837489992380142, 0.05202331766486168, -0.036243125796318054, 0.037176959216594696, -0.09029994904994965, -0.1683429628610611, -0.0007763148751109838, 0.010803398676216602, 0.03168312832713127, 0.04643981158733368, 0.03667012229561806, 0.061239324510097504, -0.01880047656595707, -0.024256091564893723, -0.2065766304731369, -0.008815872482955456, 0.12412996590137482, -0.022375915199518204, 0.0724588930606842, -0.08079487085342407, 0.13203540444374084, -0.039120372384786606, -0.02855866029858589, 0.005213533993810415, 0.031375303864479065, -0.08930839598178864, -0.12567098438739777, -0.06602675467729568, 0.061225954443216324, 0.022498471662402153, 0.0596780888736248, 0.0024417112581431866, 0.19193901121616364, -0.1333344727754593, 0.08866282552480698, 0.10110592097043991, -0.24191753566265106, -0.03025546856224537, 0.14384306967258453, -0.024540316313505173, 0.15543513000011444, -0.07628224045038223, -0.10719222575426102, 0.08085530251264572, 0.0496523417532444, -0.056598056107759476, -0.0009009919012896717, -0.09088310599327087, 0.006839253939688206, -0.13980844616889954, -0.05230165645480156, 0.15110687911510468, 0.026294633746147156, -0.037592023611068726, -0.031246718019247055, -0.0950726643204689, -0.35348814725875854, 0.024415571242570877, -0.002826953772455454, -0.06987019628286362, 0.048015810549259186, 0.035156507045030594, -0.008995379321277142, 0.004349078983068466, -0.09262323379516602, -0.04143654555082321, 0.10029355436563492, 0.04808283597230911, 0.04402211308479309, 0.019573308527469635, 0.1058531403541565, -0.12883538007736206, -0.021110432222485542, -0.041084397584199905, -0.02869381383061409, -0.050102680921554565, -0.014695894904434681, -0.07354352623224258, 0.17733292281627655, 0.06763829290866852, 0.09666602313518524, -0.1813267171382904, 0.1257096230983734, -0.02475627325475216, 0.05742272362112999, -0.035531818866729736, 0.014435730874538422, -0.11885007470846176, 0.12319590896368027, 0.0007600552635267377, 0.13899193704128265, 0.02280823327600956, -0.04473873972892761, -0.0730285570025444, -0.007154371589422226, 0.1233254000544548, 0.004622994922101498, -0.11530132591724396, 0.007921767421066761, -0.1411384791135788, -0.04554228112101555, 0.0911121591925621, -0.07906953990459442, 0.015561599284410477, 0.041121818125247955, -0.04623573645949364, 0.004239376168698072, 0.09626287966966629, -0.05426957085728645, -0.04919392615556717, -0.03332843258976936, -0.10088252276182175, 0.00408816896378994, -0.1039951965212822, -0.13126124441623688, 0.051543787121772766, -0.1563965529203415, -0.00575046194717288, -0.04191170632839203, -0.06613527983427048, 0.02011139690876007, 0.019768383353948593, -0.08437694609165192, 0.05859275907278061, -0.08150110393762589, -0.1557578295469284, -0.03541033715009689, 0.018620938062667847, 0.007860139012336731, -0.02999168075621128, 0.09800084680318832, 0.032462093979120255, 0.10065783560276031, -0.1774347871541977, -0.0032585663720965385, 0.013272988609969616, 0.06698489189147949, 0.016430236399173737, 0.11852112412452698, -0.0984061062335968, -0.029854416847229004, -0.05219494178891182, -0.06776446104049683, -0.10671226680278778, -0.009978118352591991, 0.12339237332344055, 0.081203892827034, -0.15919211506843567, -0.010960342362523079, 0.07894296199083328, -0.034127745777368546, -0.08150023221969604, 0.1462218016386032, -0.043517038226127625, 0.10362964868545532, -0.03142445161938667, 0.07441626489162445, 0.2393224835395813, -0.13294187188148499, 0.0026927399449050426, 0.11512340605258942, 0.04870401695370674, -0.024512527510523796, 0.001410747179761529, 0.06937900930643082, -0.142485573887825, 0.03395999222993851, 0.067036472260952, 0.03290358558297157, -0.059946075081825256, -0.05717241391539574, -0.03879120573401451, -0.054277822375297546, 0.11310963332653046, 0.028706887736916542, 0.012219887226819992, -0.07421029359102249, -0.07091692090034485, 0.14475229382514954, 0.13386821746826172, -0.01724834553897381, -0.005964277777820826, -0.1111484244465828, 0.002316186437383294, -0.04553208500146866, 0.027614165097475052, -0.12323106080293655, 0.024642376229166985, 0.08063367009162903, 0.01609816960990429, -0.005867816042155027, 0.048622094094753265, 0.06404208391904831, 0.02775740437209606, -0.054996591061353683, 0.023086974397301674, -0.05230040103197098, -0.002152500906959176, -0.10709255933761597, -0.07857770472764969, -0.013062533922493458, -0.010289276950061321, 0.18266798555850983, -0.14896303415298462, 0.03612537682056427, 0.10309912264347076, 0.004995338153094053, -0.010612043552100658, -0.036244068294763565, -0.07481951266527176, 0.11431097984313965, -0.012485682964324951, -0.03552611544728279, 0.04104119539260864, 0.034952957183122635, -0.03309863805770874, -0.1580384373664856, -0.09931837022304535, 0.05646728724241257, 0.13403663039207458, 0.07577018439769745, -0.06703833490610123, -0.047553446143865585, -0.020618250593543053, -0.03879788890480995, 0.056729573756456375, -0.05360506474971771, 0.023469123989343643, 0.012125842273235321, 0.06819912791252136, -0.10050064325332642, -0.03629688918590546, 0.06696230918169022, -0.011795254424214363, -0.04454666003584862, 0.11060863733291626, 0.01751997321844101, -0.07169803977012634, 0.07787928730249405, 0.05941096320748329, -0.14069285988807678, 0.08998789638280869, -0.004688691347837448, -0.023223139345645905, -0.09080317616462708, 0.17682680487632751, 0.024636752903461456, 0.12305090576410294, -0.1265403926372528, 0.10609687864780426, -0.010145329870283604, 0.014946194365620613, 0.07392288744449615, -0.20129480957984924, -0.004805627278983593, -0.0471431165933609, -0.09572996944189072, -0.05821099132299423, -0.015557871200144291, 0.0170791894197464, 0.043467629700899124, -0.005354706663638353, 0.06044595688581467, 0.142664834856987, -0.018936002627015114, -0.08370210230350494, 0.17457811534404755, -0.21770337224006653, -0.20481646060943604, -0.23560066521167755, 0.0029743367340415716, -0.11851438134908676, -0.037289947271347046, -0.05196189507842064, -0.07753796130418777, 0.03229401260614395, -0.08353889733552933, -0.040616557002067566, -0.019290516152977943, 0.003061749739572406, 0.05178794264793396, 0.011813749559223652, 0.1695105880498886, -0.0735611766576767, 0.023279182612895966, 0.046429067850112915, -0.02350066788494587, 0.10787586867809296, -0.08872182667255402, -0.03777230158448219, 0.12821447849273682, -0.009709981270134449, 0.02801087126135826, 0.01423876266926527, 0.3048967123031616, 0.010030297562479973, 0.02763240598142147, 0.09710080921649933, 0.009545043110847473, 0.06174734979867935, 0.07832355052232742, 0.015707092359662056, -0.10942601412534714, 0.06433121860027313, 0.04781096801161766, -0.09661031514406204, -0.13850480318069458, -0.047762222588062286, -0.07341255247592926, 0.006593871861696243, 0.08610724657773972, 0.07147743552923203, 0.07943940162658691, 0.07025924324989319, 0.023519491776823997, 0.11204782873392105, -0.004567211028188467, -0.005818400532007217, 0.1201663389801979, -0.020266585052013397, 0.04491112753748894, -0.030482962727546692, 0.03723101690411568, 0.07022300362586975, 0.14435730874538422, 0.08234275877475739, -0.0826135203242302, -0.0026839443016797304, 0.04897516965866089, 0.2717514932155609, -0.01714148186147213, 0.10665575414896011, -0.06918495893478394, -0.013546466827392578, 0.0014700466999784112, -0.03347883000969887, -0.07884032279253006, 0.04058695584535599, -0.02208472229540348, 0.07936602830886841, -0.017719898372888565, -0.027150988578796387, 0.07915937900543213, 0.12511037290096283, 0.15299907326698303, -0.28453946113586426, -0.11869988590478897, -0.013159751892089844, 0.11041011661291122, -0.10107292234897614, 0.0223299078643322, 0.21577490866184235, 0.008470140397548676, -0.08149175345897675, -0.03681793436408043, 0.03811902552843094, -0.01879933662712574, 0.016045289114117622, 0.12573954463005066, 0.14544515311717987, 0.005644232500344515, 0.07689831405878067, -0.2924128472805023, 0.023194577544927597, 0.05471178889274597, 0.046213310211896896, -0.04598768800497055, 0.008340165950357914, -0.04420474171638489, -0.06625235080718994, 0.029148653149604797, 0.01076830830425024, 0.1823798418045044, -0.28289252519607544, -0.06890449672937393, -0.02097264863550663, 0.1087518259882927, 0.08112131059169769, 0.04162963479757309, 0.01758452132344246, 0.056389037519693375, 0.06615620851516724, 0.03817380219697952, -0.0343405120074749, -0.09718122333288193, 0.0010788479121401906, 0.15619537234306335, -0.14491763710975647, -0.05582432448863983, -0.06452815234661102, -0.024293042719364166, 0.05778651684522629, -0.1548195481300354, -0.05276089906692505, -0.06254495680332184, 0.015085814520716667, 0.1334981471300125, -0.023490145802497864, -0.01085786335170269, -0.0201820470392704, 0.024849023669958115, -0.04709408059716225, -0.09744754433631897, 0.11661811918020248, -0.0471058264374733, -0.1306038200855255, -0.019828403368592262, 0.1378549486398697, 0.08238520473241806, -0.015439215116202831, -0.0857001468539238, -0.04882924631237984, 0.027058644220232964, -0.15197721123695374, 0.005223190877586603, 0.09445921331644058, -0.0594605952501297, 0.09092224389314651, -0.11116138100624084, 0.22619131207466125, -0.04533985257148743, 0.09459085017442703, 0.061151355504989624, 0.2978513538837433, -0.08713971078395844, 0.018928516656160355, 0.04804879426956177, -0.018178772181272507, -0.2438703179359436, 0.04353063553571701, 0.05640896409749985, 0.040471117943525314, -0.03873369097709656, -0.1723128706216812, 0.035184476524591446, 0.09765327721834183, 0.009074017405509949, 0.20277278125286102, -0.32115042209625244, -0.05401403456926346, 0.04298613592982292, 0.06709850579500198, 0.15412142872810364, -0.05032355710864067, 0.005308091174811125, 0.00299404701218009, -0.025177884846925735, 0.14879994094371796, -0.10914967209100723, 0.11319706588983536, -0.028690336272120476, 0.024565158411860466, 0.009507894515991211, -0.03613625466823578, 0.14931310713291168, 0.013264713808894157, 0.089763343334198, 0.019067833200097084, -0.04837853088974953, 0.05236007273197174, -0.08060209453105927, 0.048760365694761276, -0.055018678307533264, 0.08426039665937424, -0.0716785341501236, 0.014209864661097527, -0.0624924935400486, -0.02230324223637581, -0.07208869606256485, -0.03789118677377701, -0.098033107817173, 0.07829388231039047, -0.014118213206529617, -0.0252530500292778, -0.030577998608350754, 0.06026613339781761, 0.06182622164487839, 0.4425763189792633, -0.06215111166238785, -0.059386324137449265, 0.0778389647603035, 0.09745035320520401, -0.018147382885217667, 0.10684601962566376, -0.12679900228977203, 0.04500621557235718, 0.12372613698244095, -0.0008733426802791655, 0.12444817274808884, 0.0908789336681366, -0.11470408737659454, -0.01735471934080124, 0.03941637650132179, -0.13241484761238098, -0.072132408618927, -0.02437761053442955, -0.0033349243458360434, -0.10859358310699463, 0.02123371511697769, 0.10324949026107788, -0.028791457414627075, 0.054798197001218796, 0.03556765988469124, 0.044085126370191574, -0.13439145684242249, 0.13734261691570282, 0.03844649717211723, 0.0715968981385231, -0.09430023282766342, 0.08732116222381592, 0.023660900071263313, 0.012141572311520576, 0.04477553442120552, -0.0023568833712488413, -0.10149598866701126, 0.010059066116809845, -0.038228947669267654, -0.09548848867416382, 0.09702998399734497, -0.04586396366357803, -0.0412009172141552, -0.11459630727767944, 0.016487689688801765, 0.07441749423742294, 0.06058561056852341, 0.10161381214857101, -0.0266474150121212, 0.01351380068808794, -0.1312156319618225, 0.0554811991751194, -0.034641385078430176, 0.028874071314930916, -0.15108218789100647, 0.060451265424489975, -0.01823197305202484, 0.07453461736440659, -0.019939351826906204, -0.013813626021146774, -0.22570104897022247, 0.021174952387809753, -0.033057328313589096, 0.003854290582239628, 0.029487796127796173, 0.03671148419380188, 0.02527959644794464, 0.05630463361740112, -0.031085221096873283, 0.04344148188829422, -0.03895607963204384, -0.051280442625284195, 0.04188956692814827, 0.0007596524665132165, -0.035418447107076645, -0.046238139271736145, 0.04512936249375343, -0.11377303302288055, 0.037379465997219086, 0.037038300186395645, -0.05313664302229881, 0.07457074522972107, 0.0551333986222744, 0.020679296925663948, 0.09771078824996948, 0.05872461944818497, 0.05539431795477867, -0.05644817277789116, 0.038207992911338806, -0.013830473646521568, -0.0014212807873263955, 0.05458271503448486, 0.13561320304870605, -0.05696135014295578, -0.05802462622523308, -0.1356644332408905, -0.022950179874897003, -0.04388374090194702, 0.04029588773846626, 0.15611368417739868, 0.10005352646112442, 0.08508609980344772, -0.08406435698270798, -0.023177621886134148, -0.1434895098209381, -0.0838560089468956, 0.057367514818906784, -0.04613112285733223, -0.013442962430417538, -0.03551625832915306, 0.07219497859477997, 0.007350239437073469, 0.15389199554920197, -0.09119492769241333, -0.09884331375360489, -0.050555817782878876, -0.17701447010040283, -0.1346471607685089, -0.005777652841061354, 0.2509252429008484, 0.031908661127090454, -0.030132247135043144, -0.06317965686321259, 0.00790130253881216, 0.0689205676317215, 0.13760150969028473, 0.04632747173309326, 0.07801228016614914, -0.12434402108192444, 0.10745592415332794, 0.04290035367012024, -0.05061715841293335, 0.12014038860797882, 0.30707380175590515, -0.08199357986450195, 0.011254170909523964, -0.08713367581367493, 0.08879105001688004, 0.02187724970281124, -0.1383187174797058, 0.013609658926725388, -0.03021937422454357, -0.15772560238838196, -0.1141202449798584, 0.03438706323504448, -0.06414132565259933, -0.20812147855758667, -0.02861359901726246, -0.10091375559568405, -0.07548198103904724, 0.1385963410139084, 0.036400459706783295, -0.01720549911260605, 0.1921585202217102, -0.07791507244110107, 0.047225624322891235, -0.010112340562045574, -0.01755616068840027, -0.004292663186788559, -0.03010612167418003, -0.10022464394569397, 0.14229239523410797, 0.0026088182348757982, 0.11017946153879166, 0.0076710316352546215, 0.08859197050333023, 0.05187401920557022, -0.024778150022029877, -0.040657348930835724, -0.0020166346803307533, 0.015454727225005627, -0.06949343532323837, 0.1287740021944046, 0.06015017256140709, -0.08151866495609283, -0.08153397589921951, -0.020808378234505653, -0.08484478294849396, -0.021831484511494637, -0.1572553515434265, 0.28017404675483704, -0.03308238461613655, 0.10164105892181396, -0.009170068427920341, -0.05692457780241966, -0.10435646772384644, 0.14374536275863647, 0.120839424431324, -0.13673727214336395, -0.012003384530544281, 0.08866086602210999, -0.007145259529352188, -0.07515248656272888, 0.155784472823143, 0.09963296353816986, -0.014980965293943882, 0.02876802533864975, -0.023070234805345535, -0.03320426493883133, 0.0033824562560766935, -0.005501914769411087, -0.034212253987789154, 0.005797996185719967, 0.03244750574231148, -0.13621188700199127, -0.017982242628932, -0.06249391660094261, -0.06658949702978134, 0.1939048320055008, -0.1203932911157608, -0.08973969519138336, -0.0308162122964859, -0.09773550927639008, -0.10429038852453232, 0.02302357368171215, -0.11399734765291214, 0.059150729328393936, 0.08660008758306503, -0.05918910726904869, -0.015915945172309875, -0.07759834825992584, -0.004432562738656998, 0.01995953358709812, 0.06680567562580109, -0.012953863479197025, 0.07061396539211273, 0.10941770672798157, -0.02109832689166069, -0.05627770721912384, 0.09945710748434067, 0.005617896094918251, -0.058368053287267685, -0.14522160589694977, 0.017044585198163986, -0.029553061351180077, 0.10308431833982468, 0.024984009563922882, -0.06643787026405334, -0.03581478074193001, -0.17625966668128967, -0.011637690477073193, -0.1381247639656067, -0.08038759231567383, -0.07788003236055374, 0.10739508271217346, 0.19482040405273438, -0.05468401312828064, 0.025051061064004898, -0.03843652456998825, 0.03321219980716705, -0.04963286593556404, 0.04273268207907677, -0.005887014791369438, -0.14593636989593506, 0.05899544060230255, -0.05112109333276749, 0.004382130224257708, -0.28370237350463867, 0.001215865253470838, 0.005580263677984476, -0.033779751509428024, -0.05040258169174194, 0.1377897560596466, 0.00042426568688824773, 0.07757376879453659, -0.062357865273952484, -0.24849353730678558, -0.0690150111913681, 0.14145798981189728, 0.0006318931700661778, -0.07864443957805634 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text-generation
khanhnto/kyt-tietest-13b
[ "transformers", "safetensors", "llama", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T11:53:26+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 56, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.06061961501836777, 0.15481999516487122, -0.004844071343541145, 0.02074851468205452, 0.0983177199959755, 0.007407687604427338, 0.07119518518447876, 0.11185134947299957, -0.023851769044995308, 0.1167980208992958, 0.031993988901376724, 0.09781743586063385, 0.11217817664146423, 0.16186554729938507, 0.0015333457849919796, -0.22897611558437347, 0.049678247421979904, -0.125278040766716, -0.0294334813952446, 0.11977242678403854, 0.1422213912010193, -0.10954539477825165, 0.0752737894654274, -0.038042325526475906, -0.005828251596540213, -0.0323176346719265, -0.06205610930919647, -0.05266609415411949, 0.05311284959316254, 0.06794639676809311, 0.07308239489793777, 0.01171939354389906, 0.09106900542974472, -0.2724283039569855, 0.02348201349377632, 0.0805930644273758, -0.0006441773730330169, 0.07586129754781723, 0.04993962123990059, -0.08749990910291672, 0.07524524629116058, -0.060156844556331635, 0.1498761922121048, 0.07955671846866608, -0.09018243104219437, -0.19217631220817566, -0.07921334356069565, 0.09916994720697403, 0.1890910118818283, 0.05953684076666832, -0.026427440345287323, 0.11642678081989288, -0.08593545109033585, 0.013638701289892197, 0.06446459144353867, -0.06054406240582466, -0.055855002254247665, 0.06904532760381699, 0.08335285633802414, 0.08567540347576141, -0.12976622581481934, -0.010767064057290554, 0.015032444149255753, 0.008952446281909943, 0.08948688954114914, 0.017146794125437737, 0.1335189938545227, 0.040557652711868286, -0.13501930236816406, -0.043155476450920105, 0.09761431813240051, 0.03665134683251381, -0.04888195917010307, -0.2485782504081726, -0.023432478308677673, -0.04339504987001419, -0.03198111802339554, -0.03649339824914932, 0.043764639645814896, -0.014506848528981209, 0.07738617807626724, -0.004502781666815281, -0.0837155357003212, -0.04301247000694275, 0.07241875678300858, 0.06128999963402748, 0.02571401372551918, -0.015821760520339012, 0.0059297760017216206, 0.12327717989683151, 0.11431120336055756, -0.126715749502182, -0.052547648549079895, -0.06306339055299759, -0.08449548482894897, -0.044861067086458206, 0.030838407576084137, 0.037995077669620514, 0.045936476439237595, 0.23867325484752655, 0.007765117567032576, 0.053257301449775696, 0.04455438256263733, 0.014407169073820114, 0.06501194834709167, 0.11008983850479126, -0.05894824117422104, -0.09719445556402206, -0.028582042083144188, 0.10156717151403427, 0.007986726239323616, -0.04139331728219986, -0.05712985619902611, 0.07059531658887863, 0.018587570637464523, 0.12360043078660965, 0.08000938594341278, 0.003056557849049568, -0.0755772516131401, -0.062465377151966095, 0.17764076590538025, -0.15825673937797546, 0.04532013460993767, 0.03055616281926632, -0.0341108962893486, -0.009745313785970211, 0.012105142697691917, 0.025474950671195984, -0.021481726318597794, 0.09522198140621185, -0.05601342022418976, -0.034448131918907166, -0.11389608681201935, -0.03694311901926994, 0.030394554138183594, 0.011153047904372215, -0.02865210548043251, -0.03502652049064636, -0.08865131437778473, -0.06405586749315262, 0.09101516753435135, -0.07148737460374832, -0.04784895107150078, -0.016645915806293488, -0.07833752781152725, 0.021804187446832657, 0.01691517047584057, 0.09064167737960815, -0.0222476739436388, 0.03985358029603958, -0.0550384595990181, 0.061440225690603256, 0.11723454296588898, 0.027987057343125343, -0.05787884071469307, 0.061519939452409744, -0.2424532175064087, 0.10252492874860764, -0.07715212553739548, 0.04971238598227501, -0.15203025937080383, -0.02478341944515705, 0.03986154496669769, 0.01284773275256157, -0.008251311257481575, 0.14196595549583435, -0.21994100511074066, -0.030957341194152832, 0.16964265704154968, -0.10025953501462936, -0.08109250664710999, 0.060782887041568756, -0.05354252830147743, 0.11210215091705322, 0.04557164013385773, -0.02375967986881733, 0.05775221437215805, -0.14725260436534882, -0.011030761525034904, -0.041942402720451355, -0.0180682260543108, 0.16207332909107208, 0.0703711211681366, -0.06047816202044487, 0.07456906884908676, 0.01960151270031929, -0.014246034435927868, -0.04887177795171738, -0.02822130173444748, -0.1047162413597107, 0.01184528972953558, -0.06102835759520531, 0.018109694123268127, -0.021768750622868538, -0.09445013850927353, -0.029118487611413002, -0.17402999103069305, -0.0031633328180760145, 0.08821269869804382, -0.011630427092313766, -0.021509924903512, -0.11245372891426086, 0.009332616813480854, 0.030967719852924347, 0.0002618339203763753, -0.13677829504013062, -0.06033218279480934, 0.026970699429512024, -0.16097871959209442, 0.029791243374347687, -0.05741601809859276, 0.04530094936490059, 0.04005871340632439, -0.03433511033654213, -0.03489551320672035, 0.010874404571950436, 0.010431389324367046, -0.01894843392074108, -0.25422003865242004, -0.01882786676287651, -0.0234990194439888, 0.1751047968864441, -0.22956320643424988, 0.042598169296979904, 0.07489731162786484, 0.1460893303155899, 0.007349682506173849, -0.03550100699067116, 0.015185600146651268, -0.07262228429317474, -0.03268764168024063, -0.06316669285297394, -0.01207790058106184, -0.038400664925575256, -0.05820201337337494, 0.04906858503818512, -0.1686294972896576, -0.030321966856718063, 0.10717973858118057, 0.06342670321464539, -0.1473218947649002, -0.02780107781291008, -0.04056945815682411, -0.04624456167221069, -0.06676914542913437, -0.05461418256163597, 0.11812574416399002, 0.056411582976579666, 0.04860803112387657, -0.07140495628118515, -0.07455260306596756, 0.008036690764129162, -0.01956399530172348, -0.014917809516191483, 0.09334591031074524, 0.07554110884666443, -0.12264352291822433, 0.09177418053150177, 0.09668384492397308, 0.08576478064060211, 0.10314212739467621, -0.014663571491837502, -0.08914592862129211, -0.040637146681547165, 0.02245822176337242, 0.016187267377972603, 0.15129362046718597, -0.012961224652826786, 0.055492039769887924, 0.0358695350587368, -0.014034898020327091, 0.011105312965810299, -0.09736533463001251, 0.02655916102230549, 0.030835967510938644, -0.016302183270454407, 0.03745110332965851, -0.0447014644742012, 0.019208140671253204, 0.09039704501628876, 0.040895868092775345, 0.040978945791721344, 0.010155045427381992, -0.04354988783597946, -0.11037563532590866, 0.1787576973438263, -0.12389461696147919, -0.24818050861358643, -0.13812170922756195, 0.010281167924404144, 0.04737642779946327, -0.010411068797111511, 0.006690691225230694, -0.06616118550300598, -0.1175973042845726, -0.09878289699554443, 0.018617089837789536, 0.045352302491664886, -0.07590975612401962, -0.06842505931854248, 0.06414616107940674, 0.03875524550676346, -0.13939815759658813, 0.024007495492696762, 0.04662325978279114, -0.08205481618642807, -0.0029386086389422417, 0.0791812464594841, 0.06965780258178711, 0.17661017179489136, 0.013885351829230785, -0.023669935762882233, 0.026634456589818, 0.20819635689258575, -0.1436755359172821, 0.10975687950849533, 0.13545554876327515, -0.08767466992139816, 0.08120133727788925, 0.1998777538537979, 0.03777998685836792, -0.10680917650461197, 0.03608465939760208, 0.028374753892421722, -0.028325283899903297, -0.2502254545688629, -0.06958996504545212, 0.0019060121849179268, -0.05172049254179001, 0.07064855098724365, 0.08791537582874298, 0.09593888372182846, 0.016860228031873703, -0.09976044297218323, -0.07697858661413193, 0.046900223940610886, 0.10824491083621979, -0.00015424020239152014, -0.015208319760859013, 0.0904119610786438, -0.03033481352031231, 0.01743943803012371, 0.09215071052312851, 0.0030607767403125763, 0.17535938322544098, 0.051709048449993134, 0.17189906537532806, 0.07866133749485016, 0.06444311141967773, 0.02004685252904892, 0.007725914940237999, 0.021817529574036598, 0.017227526754140854, -0.0030957073904573917, -0.08709781616926193, -0.0034981227945536375, 0.1202581599354744, 0.049845851957798004, 0.029173865914344788, 0.012042860500514507, -0.030704669654369354, 0.08337877690792084, 0.1770893782377243, 0.0029054484330117702, -0.1893385946750641, -0.07169844210147858, 0.07795937359333038, -0.08648337423801422, -0.10729733109474182, -0.029470939189195633, 0.041069481521844864, -0.1729043871164322, 0.016882894560694695, -0.019335895776748657, 0.10788324475288391, -0.13190391659736633, -0.01772487722337246, 0.05657728388905525, 0.06932812184095383, -0.009677323512732983, 0.06694949418306351, -0.16090403497219086, 0.11770165711641312, 0.01751571334898472, 0.06636732816696167, -0.09608277678489685, 0.09618937969207764, -0.007830657996237278, 0.0041499207727611065, 0.1410749852657318, 0.010120149701833725, -0.05952107161283493, -0.09608154743909836, -0.10546442121267319, -0.009841260500252247, 0.1306990385055542, -0.14852415025234222, 0.08813067525625229, -0.02661319263279438, -0.044553373008966446, 0.003614129964262247, -0.12497276812791824, -0.13103094696998596, -0.18366187810897827, 0.05707118660211563, -0.12947207689285278, 0.04045100137591362, -0.10902881622314453, -0.045833900570869446, -0.02098964899778366, 0.20040063560009003, -0.23137451708316803, -0.06714103370904922, -0.1551055610179901, -0.08061286807060242, 0.14446212351322174, -0.046455029398202896, 0.08550118654966354, 0.0008278203313238919, 0.19068008661270142, 0.021319707855582237, -0.017237508669495583, 0.1072206199169159, -0.10052918642759323, -0.2010865956544876, -0.09273224323987961, 0.15895552933216095, 0.13766798377037048, 0.03809428587555885, -0.004381525795906782, 0.03171157464385033, -0.02098114788532257, -0.12076930701732635, 0.020226983353495598, 0.17317426204681396, 0.08982043713331223, 0.025265544652938843, -0.02972041629254818, -0.11267432570457458, -0.07061342149972916, -0.03774050623178482, 0.024755435064435005, 0.18072067201137543, -0.07222156971693039, 0.18405316770076752, 0.13775517046451569, -0.05534014105796814, -0.19904261827468872, 0.021996473893523216, 0.04293542355298996, 0.0070380112156271935, 0.0323902890086174, -0.20307663083076477, 0.09384101629257202, 0.0008334947633557022, -0.05131231248378754, 0.1379684954881668, -0.1823476254940033, -0.151598259806633, 0.06042521819472313, 0.043563615530729294, -0.19374065101146698, -0.12374074012041092, -0.08848230540752411, -0.04693066328763962, -0.15487661957740784, 0.10312657803297043, 0.0020827590487897396, 0.008401188999414444, 0.03778626397252083, 0.02252252586185932, 0.012139533646404743, -0.04198719933629036, 0.1914343535900116, -0.025891713798046112, 0.03347287327051163, -0.0790715217590332, -0.060851071029901505, 0.062408581376075745, -0.058187782764434814, 0.0755455270409584, -0.025226406753063202, 0.015947066247463226, -0.10598332434892654, -0.048235729336738586, -0.02852320298552513, 0.019321219995617867, -0.09431382268667221, -0.09348297864198685, -0.04829427972435951, 0.09367614984512329, 0.09042316675186157, -0.03652578964829445, -0.03649144619703293, -0.078715980052948, 0.038977332413196564, 0.17627815902233124, 0.18159319460391998, 0.04659178853034973, -0.07959239184856415, -0.001915142871439457, -0.014336181804537773, 0.04684065282344818, -0.22077152132987976, 0.060553863644599915, 0.04557652771472931, 0.016117896884679794, 0.11537692695856094, -0.0208132341504097, -0.16198977828025818, -0.06710557639598846, 0.061360616236925125, -0.06944561004638672, -0.17825035750865936, 0.0039279889315366745, 0.07344977557659149, -0.16578389704227448, -0.037031736224889755, 0.04200848564505577, -0.01189455483108759, -0.0403641052544117, 0.012352054007351398, 0.08063354343175888, 0.007078902795910835, 0.07699975371360779, 0.055281639099121094, 0.09124495089054108, -0.10227900743484497, 0.07410510629415512, 0.08149529248476028, -0.08644098788499832, 0.030720343813300133, 0.09573426842689514, -0.06469762325286865, -0.0346054881811142, 0.04237886518239975, 0.08354541659355164, 0.024281201884150505, -0.04682289808988571, 0.0023111123591661453, -0.09734189510345459, 0.05927345156669617, 0.11483542621135712, 0.03496333956718445, 0.011234734207391739, 0.03813567012548447, 0.04486291855573654, -0.08093374222517014, 0.11926916986703873, 0.023795632645487785, 0.020354853942990303, -0.04112942889332771, -0.040553025901317596, 0.035851649940013885, -0.026020776480436325, -0.011440055444836617, -0.035174157470464706, -0.0722682997584343, -0.014069457538425922, -0.16000694036483765, -0.0076758842915296555, -0.03660871088504791, 0.005114538595080376, 0.022510098293423653, -0.03652830421924591, 0.00792311318218708, 0.012217256240546703, -0.06868947297334671, -0.05553458258509636, -0.023233558982610703, 0.09422210603952408, -0.16494666039943695, 0.0220257006585598, 0.0823851153254509, -0.12121747434139252, 0.09289738535881042, 0.016782134771347046, 0.00412249518558383, 0.026962365955114365, -0.1545863002538681, 0.04763968288898468, -0.020152103155851364, 0.013473534025251865, 0.04222847521305084, -0.21637047827243805, -0.004404853098094463, -0.04015503451228142, -0.05566934496164322, -0.008993052877485752, -0.0319182425737381, -0.11338426172733307, 0.09645436704158783, 0.011025024577975273, -0.08443772792816162, -0.02965564839541912, 0.03353232145309448, 0.07690354436635971, -0.027447547763586044, 0.1498211771249771, -0.004663881380110979, 0.07559948414564133, -0.17581342160701752, -0.02282017655670643, -0.011197620071470737, 0.022367527708411217, -0.021871577948331833, -0.01622559316456318, 0.04623444378376007, -0.02704801969230175, 0.19120801985263824, -0.024701936170458794, 0.049393873661756516, 0.06364397704601288, 0.009232889860868454, -0.013832193799316883, 0.11151392012834549, 0.05708572641015053, 0.024334950372576714, 0.022262847051024437, 0.003451440716162324, -0.04008655622601509, -0.009981024079024792, -0.18596695363521576, 0.06803664565086365, 0.14585918188095093, 0.09060460329055786, -0.012669353745877743, 0.0707244873046875, -0.10161512345075607, -0.12005364894866943, 0.10127941519021988, -0.06415384262800217, -0.010188822634518147, -0.06542414426803589, 0.14027701318264008, 0.14953285455703735, -0.1886233240365982, 0.06583356112241745, -0.06602055579423904, -0.0566304549574852, -0.11457879096269608, -0.1930263340473175, -0.057075321674346924, -0.050602465867996216, -0.018466074019670486, -0.05384097993373871, 0.06939727067947388, 0.05750798434019089, 0.01126816775649786, 0.00868057832121849, 0.08568526059389114, -0.009656033478677273, 0.00248199631460011, 0.030120067298412323, 0.06713981181383133, 0.016768986359238625, -0.0321255661547184, 0.0179112758487463, -0.00597198773175478, 0.034156378358602524, 0.059282708913087845, 0.03608176112174988, -0.028436895459890366, 0.015559280291199684, -0.034912437200546265, -0.11309733241796494, 0.042801856994628906, -0.029640642926096916, -0.0749855786561966, 0.1347348988056183, 0.026981467381119728, 0.005015076603740454, -0.023140020668506622, 0.2503887414932251, -0.07436972856521606, -0.09334370493888855, -0.14373961091041565, 0.11701542884111404, -0.04212593287229538, 0.0635172426700592, 0.03596310690045357, -0.10810714215040207, 0.017985546961426735, 0.1320217251777649, 0.15442703664302826, -0.04732590913772583, 0.019251897931098938, 0.028577854856848717, 0.00439635943621397, -0.04075566306710243, 0.05177190154790878, 0.07100846618413925, 0.14500564336776733, -0.05157303810119629, 0.08530787378549576, 0.002609728369861841, -0.1021018698811531, -0.041973695158958435, 0.11415864527225494, -0.014296893030405045, 0.017620453611016273, -0.057136841118335724, 0.124222531914711, -0.05874236673116684, -0.23697422444820404, 0.06316976249217987, -0.0765061303973198, -0.1432730257511139, -0.024886758998036385, 0.071670763194561, -0.016632623970508575, 0.02605951391160488, 0.07167234271764755, -0.0754380151629448, 0.18880942463874817, 0.03957989811897278, -0.05233397334814072, -0.05954399332404137, 0.0744764655828476, -0.11850855499505997, 0.27879106998443604, 0.010482731275260448, 0.051307905465364456, 0.1042102724313736, -0.02021743729710579, -0.13270841538906097, 0.023401619866490364, 0.09579801559448242, -0.08917027711868286, 0.04087764397263527, 0.21448291838169098, -0.00629545608535409, 0.11935057491064072, 0.07611140608787537, -0.07468950748443604, 0.047562725841999054, -0.11468592286109924, -0.07639975845813751, -0.08699081838130951, 0.09244474768638611, -0.06785612553358078, 0.14258281886577606, 0.12599852681159973, -0.05530165135860443, 0.011584274470806122, -0.028389399871230125, 0.045467376708984375, 0.005578654818236828, 0.100032277405262, 0.011115525849163532, -0.18496567010879517, 0.024811718612909317, 0.016259413212537766, 0.10884406417608261, -0.18112654983997345, -0.09105053544044495, 0.046958595514297485, 0.0005061255069449544, -0.06443515419960022, 0.12483241409063339, 0.057313691824674606, 0.04654949903488159, -0.0451689288020134, -0.026830285787582397, -0.006042256020009518, 0.14264579117298126, -0.10707559436559677, -0.005129707511514425 ]
null
null
transformers
[speakleash.org](https://speakleash.org) ## Prompt template: ChatML ``` <|im_start|>system {system_message}<|im_end|> <|im_start|>user {prompt}<|im_end|> <|im_start|>assistant ```
{"license": "apache-2.0"}
text-generation
Nondzu/openchat-3.5-0106-speakleash-007-pl-8192-32-16-0.01
[ "transformers", "pytorch", "safetensors", "mistral", "text-generation", "conversational", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T11:56:19+00:00
[]
[]
TAGS #transformers #pytorch #safetensors #mistral #text-generation #conversational #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
URL ## Prompt template: ChatML
[ "## Prompt template: ChatML" ]
[ "TAGS\n#transformers #pytorch #safetensors #mistral #text-generation #conversational #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "## Prompt template: ChatML" ]
[ 63, 8 ]
[ "passage: TAGS\n#transformers #pytorch #safetensors #mistral #text-generation #conversational #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n## Prompt template: ChatML" ]
[ -0.03472244367003441, 0.023891953751444817, -0.006842698436230421, 0.01020998414605856, 0.11760085076093674, -0.04041735455393791, 0.12997494637966156, 0.11317834258079529, -0.021746229380369186, -0.051469963043928146, 0.10812696069478989, 0.20791468024253845, -0.016136495396494865, 0.08788542449474335, -0.07167194038629532, -0.20487642288208008, 0.10530126094818115, 0.0033248967956751585, 0.05757478252053261, 0.09610744565725327, 0.1241317018866539, -0.02645120397210121, 0.06300366669893265, 0.0045683374628424644, -0.04567689076066017, -0.002450791886076331, 0.020076945424079895, -0.09779462218284607, 0.06149734556674957, 0.037758782505989075, 0.03577270358800888, 0.0441470742225647, -0.0732150673866272, -0.19062964618206024, 0.044792771339416504, 0.015669459477066994, -0.02983354590833187, 0.030532417818903923, 0.03974676877260208, -0.07603073120117188, 0.06549765169620514, 0.007542863488197327, -0.02122681215405464, 0.08028771728277206, -0.08225320279598236, -0.0575292594730854, -0.06434590369462967, 0.053588077425956726, 0.11163181066513062, 0.12695235013961792, 0.0033946430776268244, 0.1581707000732422, -0.05090298503637314, 0.10358262807130814, 0.14717739820480347, -0.3310408890247345, -0.025229161605238914, 0.07870256900787354, 0.04109909012913704, 0.04606451839208603, -0.021084420382976532, 0.03888958692550659, 0.021485140547156334, 0.010598533786833286, 0.01889009401202202, -0.06531093269586563, -0.1896762102842331, -0.01625652052462101, -0.0710616335272789, -0.04717378690838814, 0.24890688061714172, -0.020639551803469658, 0.036113228648900986, -0.074363112449646, -0.07468121498823166, 0.07498505711555481, -0.0604369081556797, 0.05147414654493332, -0.010621041990816593, 0.08558482676744461, 0.06000654399394989, -0.07746094465255737, -0.11450371146202087, 0.004227802623063326, -0.12446698546409607, 0.0025940369814634323, 0.033998388797044754, 0.0456329807639122, -0.18189318478107452, 0.05898245424032211, 0.0925983414053917, -0.12768307328224182, -0.015465661883354187, -0.07524268329143524, 0.11796436458826065, 0.033285513520240784, -0.037649426609277725, -0.03818391263484955, 0.1839965432882309, 0.16462363302707672, -0.010431503877043724, 0.0508112907409668, -0.03881138563156128, 0.08278577774763107, -0.02992725931107998, -0.0010754561517387629, 0.05524706467986107, -0.018878642469644547, 0.13783493638038635, -0.03412843868136406, 0.09997756034135818, -0.025774234905838966, -0.127770334482193, 0.006231551058590412, 0.024860741570591927, 0.12287303805351257, 0.08265397697687149, 0.12167274206876755, -0.004999237135052681, 0.021911906078457832, 0.12713080644607544, -0.08668296039104462, -0.01148255355656147, 0.04169933497905731, 0.01909789629280567, 0.042874157428741455, 0.06654644012451172, 0.030486498028039932, -0.10979029536247253, -0.05673433840274811, -0.03082674741744995, 0.005024919286370277, -0.05925048887729645, -0.005545314401388168, 0.05031612142920494, 0.019731124863028526, -0.007691566366702318, -0.15682820975780487, -0.16809138655662537, -0.0018115448765456676, 0.07108677178621292, 0.006948959082365036, -0.1179472953081131, -0.04701223596930504, -0.045157358050346375, 0.013392235152423382, -0.07082689553499222, -0.016103733330965042, -0.08586933463811874, 0.04901248589158058, -0.08726667612791061, 0.024627776816487312, -0.1427457630634308, 0.05035853758454323, -0.13285796344280243, -0.010276627726852894, -0.10072896629571915, 0.046104565262794495, -0.06130111590027809, 0.19381386041641235, -0.031422387808561325, 0.007219519466161728, -0.0026939287781715393, 0.006756836548447609, -0.025755994021892548, 0.18529565632343292, -0.10129469633102417, -0.032191719859838486, 0.27523085474967957, -0.13136252760887146, -0.21419985592365265, 0.0921364352107048, 0.015174795873463154, 0.02454984188079834, 0.10303061455488205, 0.17585161328315735, 0.09231503307819366, -0.014980367384850979, 0.07263445109128952, 0.09798220545053482, -0.05232701078057289, -0.10388603061437607, -0.0008018108201213181, -0.060902711004018784, -0.09961524605751038, 0.07296928763389587, 0.022051896899938583, 0.07966886460781097, 0.0035404940135776997, -0.056596703827381134, -0.07386942207813263, -0.011806407012045383, -0.05435466766357422, -0.04475655406713486, 0.026558591052889824, -0.09156285971403122, -0.009259186685085297, 0.004095423500984907, 0.03335946053266525, 0.0005400220397859812, 0.03754246234893799, -0.04915487393736839, 0.057076338678598404, 0.0750562846660614, 0.06972011923789978, -0.0959993451833725, -0.08724422752857208, -0.03340189903974533, 0.028960147872567177, 0.011093705892562866, 0.013056696392595768, 0.06939594447612762, -0.016956834122538567, 0.018700011074543, -0.017401792109012604, 0.1613021194934845, 0.013942190445959568, -0.03770773857831955, -0.13130919635295868, 0.06199885159730911, -0.04211147129535675, 0.14422181248664856, -0.04517936334013939, 0.06405388563871384, 0.01659698225557804, 0.1274389773607254, -0.018233109265565872, 0.05378882214426994, 0.04808821529150009, -0.039554350078105927, -0.04347076267004013, 0.005869167856872082, 0.11977292597293854, 0.06872668862342834, -0.0649833083152771, 0.2247747778892517, -0.18292571604251862, 0.11510154604911804, 0.19543464481830597, -0.1747245490550995, 0.07840027660131454, -0.1271747499704361, -0.04971010237932205, 0.009239112958312035, 0.039071667939424515, -0.0238951425999403, 0.10954630374908447, 0.0038079190999269485, 0.1659998893737793, -0.05882764235138893, -0.04819410294294357, -0.03958866000175476, -0.08293803781270981, -0.01035058032721281, 0.08342840522527695, 0.07294470816850662, -0.10859160125255585, 0.1629076451063156, 0.20934824645519257, 0.012318146415054798, 0.12318092584609985, -0.06283117830753326, 0.02349499613046646, 0.07599407434463501, 0.06512171030044556, -0.03182416781783104, -0.04300631955265999, -0.20670892298221588, -0.0007491462747566402, 0.08067242801189423, 0.029642047360539436, 0.08054641634225845, -0.11346421390771866, -0.024085044860839844, -0.021212833002209663, -0.05113399401307106, 0.0254053995013237, 0.05339444428682327, 0.038314033299684525, 0.10966023802757263, -0.06601585447788239, -0.06152975559234619, 0.08758854120969772, -0.003923157230019569, -0.1043560653924942, 0.14885997772216797, -0.1640634983778, -0.272233247756958, -0.0878218486905098, -0.08258505910634995, -0.1028638556599617, -0.009608740918338299, 0.1638648509979248, -0.0723608136177063, -0.03657447174191475, -0.09683176875114441, -0.022053496912121773, 0.05113006755709648, -0.015222698450088501, -0.01716049574315548, 0.03587259352207184, -0.010146040469408035, -0.15157729387283325, -0.03943522647023201, 0.006050482392311096, -0.07524069398641586, 0.1105433776974678, -0.08585170656442642, 0.06619702279567719, 0.12354830652475357, 0.042836349457502365, 0.05645964294672012, -0.024540970101952553, 0.175403892993927, -0.005141668487340212, 0.03068096563220024, 0.23044031858444214, 0.0069938660599291325, 0.07544563710689545, 0.1707070916891098, 0.0027733384631574154, -0.060092825442552567, 0.036353111267089844, -0.031293969601392746, -0.04311215132474899, -0.21860551834106445, -0.1335398405790329, -0.11716946959495544, 0.060062091797590256, -0.002308873226866126, 0.08672311902046204, 0.07726068794727325, 0.06781262904405594, -0.08528144657611847, -0.07774389535188675, 0.12441809475421906, 0.09017723053693771, 0.1906111240386963, -0.039270270615816116, 0.10611841827630997, -0.09316065162420273, -0.07659590989351273, 0.08863896876573563, 0.04896317794919014, 0.14675316214561462, 0.05206943675875664, 0.0734984278678894, 0.08319205790758133, 0.11854025721549988, 0.05800757184624672, 0.07368090003728867, -0.011917222291231155, -0.003864779369905591, -0.03830859437584877, -0.0801183432340622, -0.05030984431505203, 0.00946870818734169, -0.10326465219259262, -0.06955686211585999, -0.04878157004714012, -0.03384675085544586, 0.13767234981060028, 0.17405633628368378, 0.016256971284747124, -0.15093214809894562, 0.005813355091959238, 0.09379984438419342, 0.017243260517716408, -0.0657867044210434, 0.09690235555171967, -0.0052035353146493435, -0.07617907226085663, 0.11250621825456619, -0.009812409058213234, 0.1532137095928192, -0.00801061000674963, 0.05300867557525635, -0.09967135637998581, -0.007893241941928864, 0.03535677492618561, 0.12027328461408615, -0.2802145183086395, 0.1606258600950241, -0.004767285659909248, 0.0196192879229784, -0.09393224120140076, 0.016093779355287552, 0.06149817258119583, 0.19478507339954376, 0.08491697162389755, 0.010192721150815487, -0.0860966145992279, 0.05115203931927681, -0.09184540808200836, 0.061767298728227615, 0.017129965126514435, 0.008128983899950981, -0.023279644548892975, -0.06665555387735367, -0.04142194613814354, -0.00943826511502266, -0.04842640459537506, -0.09551718086004257, -0.18440602719783783, 0.02471638098359108, 0.1645750105381012, 0.05342530459165573, -0.049370620399713516, 0.011173712089657784, -0.04890655353665352, 0.22004930675029755, 0.025507904589176178, -0.11670814454555511, -0.0916900485754013, -0.15769170224666595, -0.0011670859530568123, -0.05401939898729324, 0.04462385177612305, -0.09643088281154633, 0.05013102665543556, -0.06497166305780411, -0.18919707834720612, 0.08416904509067535, -0.14209187030792236, 0.006538149435073137, -0.016473911702632904, 0.18144004046916962, -0.07879868894815445, -0.015046990476548672, 0.06429994106292725, 0.0018809974426403642, -0.11446904391050339, -0.12915337085723877, -0.0438377670943737, 0.10738620907068253, 0.037749603390693665, 0.0006246614502742887, -0.15443728864192963, -0.2107686996459961, -0.01851336658000946, -0.02357359044253826, 0.2570290267467499, 0.20060671865940094, -0.043325040489435196, 0.1351471245288849, 0.16886864602565765, -0.06728433817625046, -0.28761184215545654, -0.11752693355083466, -0.1405085027217865, -0.11188175529241562, -0.05332468822598457, -0.10960901528596878, 0.09022988379001617, 0.015061198733747005, -0.054475001990795135, 0.12653031945228577, -0.26910263299942017, -0.08127672970294952, 0.1823030412197113, 0.038570865988731384, 0.2856511175632477, -0.16864466667175293, -0.09596693515777588, -0.06635694205760956, -0.1721120923757553, 0.1600758582353592, -0.20435254275798798, 0.04526503384113312, 0.027990475296974182, 0.04590025916695595, 0.01819334179162979, -0.05219075828790665, 0.08441507816314697, -0.031297117471694946, 0.015915192663669586, -0.13419489562511444, 0.06989585608243942, 0.036259010434150696, -0.020400501787662506, 0.12148439139127731, -0.20609311759471893, 0.043997202068567276, -0.0695684403181076, -0.002111750887706876, -0.07778850197792053, 0.09986109286546707, -0.002724122256040573, -0.04966839402914047, -0.0026501109823584557, -0.08253727108240128, 0.05410848557949066, 0.03601049259305, 0.17452634871006012, -0.05753099545836449, 0.162754625082016, 0.15068385004997253, 0.09756841510534286, -0.23380890488624573, -0.011596056632697582, -0.056106630712747574, -0.05936157703399658, 0.05675468593835831, -0.05949937179684639, 0.04133286327123642, 0.06089106202125549, -0.05850420147180557, 0.07595448940992355, 0.054762426763772964, 0.015736985951662064, -0.03474748507142067, 0.09665421396493912, -0.22730906307697296, -0.08727091550827026, -0.03606398031115532, 0.21458154916763306, 0.040444109588861465, 0.10321010649204254, 0.1645951271057129, -0.005665912292897701, -0.030829932540655136, 0.015919407829642296, 0.0504261776804924, -0.017036600038409233, 0.004334242083132267, -0.004885916598141193, 0.00025770365027710795, -0.1361083686351776, 0.07503681629896164, 0.01039182674139738, -0.17590974271297455, 0.0037808315828442574, 0.19542557001113892, -0.17414651811122894, -0.1667490154504776, 0.0006394648808054626, 0.1292378008365631, -0.04253225773572922, -0.06425780802965164, -0.06785480678081512, -0.1686190962791443, 0.03937526419758797, 0.0786084458231926, 0.064449243247509, 0.015712374821305275, 0.05066496878862381, -0.0555165596306324, 0.0144409304484725, 0.05834547057747841, -0.04828634858131409, -0.00009434163803234696, -0.05397256091237068, 0.01484057866036892, 0.02379005402326584, 0.05550725385546684, -0.048113420605659485, -0.04379328340291977, -0.09111691266298294, 0.010972408577799797, -0.12623032927513123, -0.0015723315300419927, -0.12664106488227844, -0.009711933322250843, -0.004643960855901241, -0.03362764045596123, -0.021316925063729286, 0.005151595454663038, -0.09254435449838638, -0.03746224194765091, -0.05795317515730858, 0.06033564731478691, -0.1413092315196991, -0.02562970481812954, 0.059436798095703125, -0.036412667483091354, 0.1268213540315628, 0.12485549598932266, -0.13770468533039093, 0.03642386198043823, -0.21519383788108826, -0.1166650801897049, 0.0941477045416832, 0.033416010439395905, 0.028822381049394608, 0.01377035304903984, -0.02174905315041542, 0.12948906421661377, 0.020108992233872414, 0.022456001490354538, 0.0953732430934906, -0.12798550724983215, -0.02103249728679657, 0.013376658782362938, -0.11268869042396545, -0.014207572676241398, -0.10137713700532913, 0.06888648122549057, 0.010814650915563107, 0.18115171790122986, -0.08676331490278244, 0.056440092623233795, -0.08956702053546906, 0.013499835506081581, 0.011984744109213352, -0.15105287730693817, -0.14043043553829193, -0.032108407467603683, 0.009908050298690796, -0.01681094989180565, 0.16952158510684967, 0.007149412762373686, -0.06916031986474991, 0.08162687718868256, 0.055776819586753845, 0.06373707950115204, 0.014940556138753891, 0.18883083760738373, 0.03549272194504738, -0.004663428291678429, -0.06373821198940277, -0.00005594859612756409, 0.051633320748806, -0.1004328802227974, 0.11348365992307663, 0.10192516446113586, 0.04578046873211861, 0.08096881210803986, 0.02867155149579048, 0.02356977015733719, -0.07091107964515686, -0.12013131380081177, 0.007595667149871588, 0.10986984521150589, -0.04829563945531845, 0.17961812019348145, 0.22523026168346405, -0.011640188284218311, 0.011865321546792984, -0.035849206149578094, -0.012894649058580399, -0.14277960360050201, -0.14031748473644257, -0.08166182786226273, -0.14026305079460144, -0.027467170730233192, -0.09341810643672943, 0.022035181522369385, 0.015183218754827976, 0.04102623090147972, -0.022557130083441734, 0.0924147292971611, 0.02499929443001747, -0.058323249220848083, 0.01693965308368206, -0.04288134351372719, 0.039559297263622284, -0.033516041934490204, -0.034352175891399384, -0.03183130547404289, -0.02055416814982891, -0.04819624498486519, 0.0968443900346756, 0.027501503005623817, 0.06624254584312439, -0.10947329550981522, -0.0706019178032875, -0.043025385588407516, 0.04980471357703209, -0.0006259779329411685, 0.15133151412010193, 0.06091354414820671, -0.026026010513305664, 0.07950414717197418, 0.20380114018917084, -0.059531547129154205, -0.19268999993801117, -0.05584811419248581, 0.1454494148492813, 0.014626928605139256, 0.0453609898686409, -0.00546625442802906, 0.0038139645475894213, -0.05531305819749832, 0.3065798580646515, 0.31358349323272705, -0.06771434843540192, 0.03618922457098961, -0.08791257441043854, 0.032760586589574814, 0.04857373982667923, 0.1140943244099617, 0.16024154424667358, 0.2521347403526306, -0.04883401095867157, 0.0025793754030019045, -0.029588596895337105, -0.0010812242981046438, -0.22621946036815643, 0.054093003273010254, -0.032876383513212204, -0.07189284265041351, 0.039467886090278625, 0.09583193808794022, -0.15013349056243896, 0.10556945949792862, -0.17705722153186798, -0.1300133764743805, -0.04671036824584007, -0.042708948254585266, 0.21597839891910553, 0.03042522445321083, 0.05383376404643059, -0.013944942504167557, -0.03706449270248413, 0.14402087032794952, -0.027994951233267784, -0.18295593559741974, -0.0834088996052742, 0.06271199882030487, -0.08305962383747101, 0.1258126050233841, 0.02379559725522995, 0.03737504407763481, 0.07212825864553452, 0.01642986387014389, -0.09732652455568314, 0.0998191311955452, 0.024097755551338196, -0.03818335384130478, 0.054309360682964325, -0.08970817178487778, -0.03284549340605736, -0.04034988582134247, 0.04772837460041046, -0.11027922481298447, 0.0499732606112957, 0.05028507485985756, -0.05967448651790619, -0.03165189176797867, 0.07266203314065933, -0.070427305996418, 0.029145654290914536, 0.0439719632267952, -0.046992890536785126, -0.02094745635986328, -0.04595399275422096, -0.026465605944395065, -0.021382661536335945, -0.183821439743042, -0.04785753786563873, -0.10607144236564636, -0.04563495144248009, 0.0792345404624939, 0.03732764720916748, -0.18329186737537384, -0.023517029359936714, -0.12436024844646454, 0.03318127617239952, -0.165377676486969, 0.05208469554781914, 0.1332733929157257, 0.010213613510131836, -0.004215145483613014, -0.0046709394082427025, 0.04668353125452995, 0.07541350275278091, -0.05099055543541908, -0.08162723481655121 ]
null
null
transformers
dict( name="igenius-llama-2b", # official hf_config=dict(org="Igenius", name="igenius-llama-2b"), block_size=2048, vocab_size=32000, padding_multiple=64, n_layer=22, n_head=64, n_embd=2560, rotary_percentage=1.0, parallel_residual=False, bias=False, _norm_class="RMSNorm", norm_eps=1e-5, _mlp_class="LLaMAMLP", intermediate_size=8960, n_query_groups=8, ), Trained on wikipedia and editorial corpus for about 2 epochs
{}
text-generation
iGenius-AI-Team/Italia-2B-ckpt-61B-wikiEditorial
[ "transformers", "safetensors", "llama", "text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T11:57:16+00:00
[]
[]
TAGS #transformers #safetensors #llama #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
dict( name="igenius-llama-2b", # official hf_config=dict(org="Igenius", name="igenius-llama-2b"), block_size=2048, vocab_size=32000, padding_multiple=64, n_layer=22, n_head=64, n_embd=2560, rotary_percentage=1.0, parallel_residual=False, bias=False, _norm_class="RMSNorm", norm_eps=1e-5, _mlp_class="LLaMAMLP", intermediate_size=8960, n_query_groups=8, ), Trained on wikipedia and editorial corpus for about 2 epochs
[ "# official\n hf_config=dict(org=\"Igenius\", name=\"igenius-llama-2b\"),\n block_size=2048,\n vocab_size=32000,\n padding_multiple=64,\n n_layer=22,\n n_head=64,\n n_embd=2560,\n rotary_percentage=1.0,\n parallel_residual=False,\n bias=False,\n _norm_class=\"RMSNorm\",\n norm_eps=1e-5,\n _mlp_class=\"LLaMAMLP\",\n intermediate_size=8960,\n n_query_groups=8,\n ),\n\n Trained on wikipedia and editorial corpus for about 2 epochs" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# official\n hf_config=dict(org=\"Igenius\", name=\"igenius-llama-2b\"),\n block_size=2048,\n vocab_size=32000,\n padding_multiple=64,\n n_layer=22,\n n_head=64,\n n_embd=2560,\n rotary_percentage=1.0,\n parallel_residual=False,\n bias=False,\n _norm_class=\"RMSNorm\",\n norm_eps=1e-5,\n _mlp_class=\"LLaMAMLP\",\n intermediate_size=8960,\n n_query_groups=8,\n ),\n\n Trained on wikipedia and editorial corpus for about 2 epochs" ]
[ 47, 154 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# official\n hf_config=dict(org=\"Igenius\", name=\"igenius-llama-2b\"),\n block_size=2048,\n vocab_size=32000,\n padding_multiple=64,\n n_layer=22,\n n_head=64,\n n_embd=2560,\n rotary_percentage=1.0,\n parallel_residual=False,\n bias=False,\n _norm_class=\"RMSNorm\",\n norm_eps=1e-5,\n _mlp_class=\"LLaMAMLP\",\n intermediate_size=8960,\n n_query_groups=8,\n ),\n\n Trained on wikipedia and editorial corpus for about 2 epochs" ]
[ -0.07190508395433426, -0.01878994144499302, -0.004910103511065245, 0.05577511712908745, 0.054755788296461105, 0.040023669600486755, 0.16387230157852173, 0.10844168066978455, 0.10825592279434204, 0.023975398391485214, 0.12223333865404129, 0.04799371212720871, 0.019963720813393593, -0.037960946559906006, -0.11810975521802902, -0.1636403203010559, 0.03769034519791603, 0.03484080359339714, 0.007485548034310341, 0.022809863090515137, 0.09199229627847672, -0.05182293802499771, 0.07410748302936554, -0.011346837505698204, -0.08185502886772156, 0.05250249430537224, 0.07350291311740875, -0.06023967266082764, 0.13775482773780823, 0.08884280174970627, 0.08800551295280457, 0.04704088345170021, 0.000757943547796458, -0.1635073721408844, 0.02901618368923664, 0.05284547433257103, 0.000014089738215261605, 0.06785479187965393, 0.0740169808268547, 0.011994754895567894, 0.06024203076958656, -0.11810269951820374, -0.050490379333496094, 0.03920072689652443, -0.09628812968730927, -0.15516915917396545, -0.03713681176304817, 0.0015106817008927464, 0.013361960649490356, 0.03314605727791786, -0.01940503530204296, 0.052703261375427246, -0.08914011716842651, 0.07507307827472687, 0.26021435856819153, -0.26147130131721497, 0.021307803690433502, 0.05440377816557884, -0.04114065319299698, 0.043298061937093735, -0.03454894572496414, 0.02872411534190178, 0.08661985397338867, 0.02498025633394718, 0.045234791934490204, -0.11497801542282104, -0.03965803608298302, -0.015016976743936539, -0.04135666787624359, 0.01210017129778862, 0.32111579179763794, 0.06420081853866577, -0.020301342010498047, -0.009516377933323383, -0.07772620767354965, 0.04258239269256592, 0.011812313459813595, 0.039270248264074326, -0.013100028969347477, -0.025578219443559647, 0.07697827368974686, -0.008623802103102207, -0.078026182949543, -0.03237653151154518, -0.11501049250364304, 0.18063367903232574, 0.05446742847561836, 0.013634700328111649, -0.05147797241806984, 0.05031120404601097, -0.08812708407640457, -0.13295868039131165, -0.023081673309206963, -0.026509415358304977, -0.02540338970720768, -0.00155250180978328, -0.06398019194602966, 0.0014342456124722958, 0.11346518993377686, 0.07869912683963776, -0.15833935141563416, 0.03460901975631714, 0.018716495484113693, 0.07283860445022583, 0.020500659942626953, -0.047307223081588745, -0.11092288047075272, 0.027973812073469162, 0.12724748253822327, 0.04666760191321373, 0.08986597508192062, -0.024018390104174614, -0.07895563542842865, -0.10932140797376633, 0.08727719634771347, 0.02392647974193096, -0.0677664652466774, 0.07644523680210114, -0.010784412734210491, 0.008572020567953587, -0.00026629058993421495, -0.12787379324436188, -0.007653425447642803, 0.034896086901426315, -0.02440963312983513, 0.06943047791719437, -0.0023126869928091764, -0.013365311548113823, -0.09937559813261032, 0.030528537929058075, -0.08639556169509888, 0.00041358411544933915, -0.014079052023589611, -0.1286993771791458, 0.045835141092538834, -0.036586955189704895, 0.020775405690073967, -0.16201409697532654, -0.11570242047309875, 0.006179082673043013, -0.006539695896208286, -0.041887134313583374, -0.02000259794294834, -0.009405899792909622, -0.04043710604310036, -0.001645452226512134, -0.02835703082382679, 0.03334908187389374, -0.07417333871126175, 0.057653989642858505, 0.02016773447394371, 0.11197246611118317, -0.17134669423103333, 0.01827634498476982, -0.07057598233222961, 0.06646087765693665, -0.08059278130531311, 0.018552206456661224, 0.010243060998618603, 0.012278000824153423, -0.022844849154353142, -0.030048903077840805, -0.08974740654230118, 0.060354363173246384, 0.03128131851553917, 0.11592389643192291, -0.1348734200000763, -0.023931600153446198, 0.16927644610404968, -0.06795245409011841, -0.13211292028427124, 0.19556555151939392, -0.034115981310606, -0.04002027586102486, 0.07976150512695312, 0.10193010419607162, -0.01771828718483448, -0.018450824543833733, -0.010751520283520222, 0.02691604569554329, 0.03834102302789688, -0.05600257217884064, 0.011627474799752235, 0.00851280428469181, -0.09355029463768005, 0.03639237582683563, 0.030823323875665665, 0.0347602516412735, -0.008345689624547958, -0.01129230298101902, -0.07267439365386963, -0.028780071064829826, -0.03813350573182106, -0.023128801956772804, 0.024054883047938347, -0.11029589921236038, -0.02508651651442051, -0.10585575550794601, 0.05324765294790268, -0.0703001543879509, 0.018088646233081818, 0.005661434028297663, 0.13211679458618164, -0.052834752947092056, 0.05221791937947273, -0.17056234180927277, -0.007685048505663872, 0.010349215939640999, 0.04987869784235954, 0.038803692907094955, 0.17617347836494446, -0.00608420092612505, 0.0038112341426312923, -0.05614587664604187, 0.011272844858467579, 0.07327735424041748, -0.002797202905640006, -0.0551651194691658, -0.13296829164028168, 0.07022448629140854, -0.023908481001853943, 0.10234713554382324, -0.23848183453083038, 0.031799666583538055, -0.0038336431607604027, 0.10962817817926407, 0.018062729388475418, 0.08511830121278763, -0.004809688776731491, 0.0452071875333786, -0.11646465957164764, 0.0047152903862297535, 0.04155288264155388, -0.011667035520076752, -0.04404916986823082, 0.016560332849621773, -0.17150048911571503, 0.10577190667390823, 0.13130098581314087, -0.1911037713289261, -0.014913820661604404, -0.03493790328502655, 0.012272053398191929, 0.010781003162264824, -0.10312503576278687, -0.09350565820932388, 0.1105627492070198, -0.0031234445050358772, 0.14604946970939636, -0.13285312056541443, -0.030522216111421585, 0.014859674498438835, -0.09758914262056351, 0.00997763592749834, 0.09151197224855423, 0.10783644765615463, -0.15628771483898163, 0.07509174942970276, 0.16561827063560486, -0.16696251928806305, 0.1792369782924652, 0.01571859046816826, -0.07054818421602249, 0.03036123886704445, 0.052053019404411316, -0.012625540606677532, 0.08500054478645325, -0.1055283173918724, -0.016787251457571983, 0.05292509123682976, 0.03316647186875343, 0.019522061571478844, -0.17434954643249512, 0.003866129321977496, -0.042353760451078415, -0.04669025540351868, 0.0005641707684844732, 0.019358329474925995, 0.054398611187934875, 0.18037374317646027, -0.046353358775377274, -0.10390643030405045, 0.031114010140299797, -0.029967278242111206, -0.1210167184472084, 0.21204346418380737, -0.03402900695800781, -0.23961098492145538, -0.10049393773078918, -0.050547339022159576, -0.1387968808412552, 0.042797114700078964, 0.033005211502313614, -0.07143253087997437, -0.05690213665366173, -0.10585886240005493, 0.04741769656538963, 0.034413181245326996, 0.027735548093914986, -0.04866205155849457, 0.0652083232998848, 0.06745262444019318, -0.0797261968255043, -0.05540602654218674, -0.034175705164670944, 0.03978538140654564, 0.11304309964179993, -0.042692556977272034, 0.06227654218673706, 0.09926693886518478, -0.018284400925040245, -0.0007857158198021352, 0.017495185136795044, 0.13185201585292816, -0.012111209332942963, 0.06224188581109047, 0.15933233499526978, -0.0074903336353600025, 0.03326433151960373, 0.19186855852603912, 0.015846244990825653, -0.0836244523525238, 0.021982882171869278, 0.06955841928720474, -0.05831141024827957, -0.18679969012737274, -0.03942467272281647, -0.10776547342538834, -0.04033524543046951, 0.08806890994310379, 0.03642118349671364, 0.028620723634958267, 0.08041997998952866, -0.06997310370206833, 0.14378905296325684, 0.0112075824290514, 0.03887222707271576, 0.23440496623516083, 0.01673092506825924, 0.09996335953474045, -0.05218350887298584, -0.09986709803342819, 0.04474976658821106, 0.0682775005698204, 0.18442317843437195, -0.04087809473276138, 0.12801659107208252, 0.025794625282287598, 0.0324513278901577, 0.07467089593410492, 0.13589325547218323, 0.010416917502880096, -0.030683627352118492, -0.02595740742981434, -0.036375440657138824, -0.024218324571847916, 0.046308476477861404, -0.08065669238567352, 0.01066066138446331, -0.06145723909139633, 0.0564790815114975, 0.05588892102241516, 0.1313692033290863, 0.1523449420928955, -0.2854821979999542, 0.030111923813819885, 0.0667819082736969, -0.002452507149428129, -0.08074075728654861, 0.04672810062766075, 0.16041816771030426, -0.003055271226912737, 0.08651284128427505, 0.020758122205734253, 0.06219793111085892, -0.1527608036994934, 0.07313933223485947, -0.08652354031801224, 0.10632564127445221, -0.02002490870654583, 0.0870528593659401, -0.2563360929489136, 0.21023684740066528, 0.045920126140117645, 0.01250841747969389, -0.10618730634450912, -0.0047632972709834576, 0.05019649118185043, 0.10885664820671082, 0.04411514103412628, 0.01732567884027958, -0.07131510972976685, -0.12531425058841705, -0.021642806008458138, 0.0704297348856926, 0.03224337100982666, 0.033204153180122375, 0.09542211145162582, -0.010211843065917492, 0.006283220369368792, 0.03073853999376297, -0.020345894619822502, -0.16228307783603668, -0.11570488661527634, 0.027235407382249832, 0.14517012238502502, -0.0588027685880661, -0.04887791723012924, -0.02564798854291439, -0.11513034254312515, 0.21112622320652008, -0.07297822088003159, -0.097478486597538, -0.08727190643548965, 0.11038712412118912, -0.009089983068406582, -0.09656308591365814, -0.09417080134153366, -0.04520378261804581, 0.10991043597459793, -0.0575655996799469, -0.09362958371639252, 0.10766596347093582, -0.056290268898010254, -0.09400805085897446, -0.07204602658748627, 0.13849808275699615, -0.04597809910774231, 0.018521226942539215, 0.05450235307216644, -0.033740103244781494, 0.00909662339836359, -0.11898381263017654, 0.03351826220750809, 0.009407985024154186, 0.044807277619838715, 0.0662178099155426, -0.14265786111354828, -0.028651688247919083, -0.025677138939499855, 0.024995071813464165, 0.16710399091243744, 0.28038349747657776, -0.014976360835134983, -0.06947953999042511, 0.1762012392282486, -0.04363013058900833, -0.21284684538841248, -0.031379349529743195, -0.07303189486265182, 0.02347530424594879, -0.060288071632385254, -0.13361674547195435, 0.1383446455001831, 0.16780665516853333, 0.01885388232767582, 0.17917433381080627, -0.0724179744720459, -0.1197662204504013, 0.1464061141014099, 0.09273362159729004, 0.30436021089553833, -0.11843639612197876, -0.029313519597053528, -0.12457980960607529, -0.05203952640295029, 0.1373414397239685, -0.16159804165363312, 0.08900656551122665, -0.03569941595196724, 0.04259537532925606, 0.021276503801345825, -0.023482546210289, 0.13798083364963531, 0.010073920711874962, 0.11601753532886505, -0.07893326878547668, 0.002976308111101389, 0.052407559007406235, -0.054414864629507065, 0.10852286219596863, -0.16438022255897522, -0.021571006625890732, -0.07401665300130844, -0.026488199830055237, -0.04699082300066948, 0.06294673681259155, -0.024083251133561134, -0.04981748014688492, -0.053451333194971085, -0.02538822405040264, -0.021666940301656723, 0.0025818217545747757, 0.15346574783325195, -0.05178995430469513, 0.09804071485996246, 0.1287342756986618, 0.1662895530462265, -0.21900007128715515, 0.07861374318599701, 0.04082530736923218, -0.022282244637608528, 0.09034620225429535, -0.09837797284126282, 0.08434242010116577, 0.08110569417476654, -0.001337085966952145, 0.1321410834789276, 0.05896605923771858, -0.04676109552383423, -0.0029767598025500774, 0.07905721664428711, -0.20714642107486725, 0.007280940655618906, -0.019909849390387535, -0.032417744398117065, -0.06315863132476807, 0.01682968996465206, 0.16957136988639832, -0.005717227701097727, 0.01897440291941166, 0.005263361148536205, 0.02203400991857052, 0.00418925192207098, 0.11987490206956863, 0.0788741484284401, 0.07098331302404404, -0.08927273750305176, 0.06310495734214783, 0.02984342724084854, -0.045196760445833206, 0.022387731820344925, 0.08591707050800323, -0.12585248053073883, -0.09019605070352554, -0.0969119668006897, 0.14617937803268433, -0.03666391968727112, -0.1152532622218132, -0.1547359973192215, -0.21025200188159943, 0.02916918322443962, 0.05862890183925629, 0.061706237494945526, -0.0014273867709562182, 0.015971409156918526, -0.1125924289226532, -0.09086018055677414, 0.09037990123033524, -0.027303947135806084, 0.073027603328228, -0.12639649212360382, -0.02398568205535412, -0.026080064475536346, 0.051679082214832306, -0.034785956144332886, 0.013843455351889133, -0.13015970587730408, -0.016713079065084457, -0.24671310186386108, 0.05271169915795326, -0.09315968304872513, -0.007211963180452585, 0.015139122493565083, 0.0025157975032925606, -0.05939893797039986, -0.012711886316537857, -0.06886938214302063, -0.021959036588668823, 0.0033404554706066847, 0.05124472826719284, -0.05510453134775162, -0.06945233792066574, -0.008200349286198616, -0.0822974443435669, 0.07858256995677948, -0.016971459612250328, -0.08318586647510529, 0.038537196815013885, -0.11638815701007843, -0.02353491634130478, 0.12915010750293732, 0.07007530331611633, 0.04685012996196747, -0.11789758503437042, 0.021770738065242767, 0.07749694585800171, 0.06863964349031448, 0.02349265106022358, -0.03673627972602844, -0.07386288046836853, 0.03717617690563202, -0.08315706253051758, -0.07013904303312302, -0.04588889330625534, 0.011046474799513817, 0.11765485256910324, 0.03481849282979965, 0.1632360816001892, -0.07022794336080551, -0.014337516389787197, -0.16015462577342987, 0.0388287715613842, 0.016455089673399925, -0.12907831370830536, 0.028585389256477356, -0.025299735367298126, 0.06779185682535172, -0.011500036343932152, 0.23372717201709747, -0.024290049448609352, -0.0199752077460289, 0.02626153826713562, 0.017722979187965393, 0.006882328540086746, 0.020136337727308273, 0.1484006643295288, 0.017131444066762924, -0.06164514645934105, -0.08678418397903442, 0.0055732931941747665, 0.17211224138736725, 0.14693664014339447, 0.23280440270900726, 0.08550719916820526, 0.0191184151917696, 0.10406982898712158, 0.0516359880566597, -0.08173377811908722, 0.09545222669839859, 0.10422004014253616, -0.0031698979437351227, 0.07553735375404358, 0.03183196112513542, -0.0009708486031740904, 0.23979683220386505, -0.06120961904525757, -0.04452887177467346, -0.05075105279684067, -0.07808027416467667, -0.14885705709457397, -0.18021775782108307, -0.11712817847728729, -0.06147973611950874, -0.03436307981610298, -0.12247179448604584, -0.06891538202762604, 0.11821077764034271, 0.04921554774045944, -0.014075610786676407, 0.0280303992331028, 0.05938317999243736, -0.07795175909996033, 0.06709150224924088, 0.01399727538228035, 0.016214638948440552, 0.009058129973709583, -0.07714362442493439, 0.041088398545980453, -0.05264025554060936, -0.00843115709722042, 0.027111276984214783, 0.05036330223083496, 0.02610931359231472, -0.09557268768548965, -0.07081273198127747, -0.020217234268784523, 0.02346736192703247, 0.022202230989933014, 0.12011086195707321, 0.0687357559800148, -0.06091325357556343, 0.03154725581407547, 0.05880904942750931, -0.028790494427084923, -0.026682408526539803, -0.09686750918626785, 0.11986863613128662, -0.036390457302331924, 0.06602004170417786, 0.001250058296136558, -0.08890005201101303, 0.02915855497121811, 0.19175489246845245, 0.19594132900238037, -0.06453771144151688, 0.030944377183914185, -0.08253760635852814, -0.0021440659184008837, -0.003486296394839883, 0.08590169996023178, 0.08634425699710846, 0.1167943999171257, -0.008843498304486275, -0.07078296691179276, -0.01998923160135746, 0.02748214453458786, -0.12672549486160278, 0.05017545446753502, 0.05139826983213425, 0.0013510226272046566, -0.05944932624697685, 0.008817790076136589, -0.05995965376496315, -0.026362575590610504, -0.002628806745633483, -0.04231083020567894, -0.08398072421550751, -0.009309395216405392, -0.058641109615564346, -0.028366712853312492, 0.05968468263745308, -0.11706547439098358, -0.040585778653621674, 0.062323398888111115, -0.0318397581577301, -0.09826154261827469, -0.021167505532503128, 0.029419919475913048, 0.049708373844623566, 0.025712862610816956, 0.011540600098669529, 0.1330510675907135, 0.10766050964593887, -0.008365500718355179, -0.07186082005500793, 0.13434283435344696, 0.014992471784353256, 0.0437740683555603, 0.04733600094914436, 0.0792958065867424, 0.03048110194504261, 0.00436555128544569, 0.12938369810581207, -0.04385046660900116, -0.0027759575750678778, -0.04167953133583069, -0.10315617173910141, -0.0853099450469017, 0.009808641858398914, -0.06218179315328598, 0.08804234862327576, 0.10150730609893799, -0.03396671265363693, -0.02379276417195797, -0.019040917977690697, -0.011396345682442188, 0.044770073145627975, -0.10463874787092209, -0.01686839386820793, -0.1729145646095276, 0.03817743808031082, 0.08317930996417999, -0.01083395816385746, -0.24548034369945526, -0.012312780134379864, -0.06341937184333801, -0.029280802235007286, -0.06634114682674408, 0.039696380496025085, 0.11485285311937332, 0.07192789763212204, -0.07303790003061295, -0.17054779827594757, -0.029865318909287453, 0.10686465352773666, -0.07543756067752838, -0.13694648444652557 ]
null
null
stable-baselines3
# **DQN** Agent playing **SpaceInvadersNoFrameskip-v4** This is a trained model of a **DQN** agent playing **SpaceInvadersNoFrameskip-v4** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3) and the [RL Zoo](https://github.com/DLR-RM/rl-baselines3-zoo). The RL Zoo is a training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included. ## Usage (with SB3 RL Zoo) RL Zoo: https://github.com/DLR-RM/rl-baselines3-zoo<br/> SB3: https://github.com/DLR-RM/stable-baselines3<br/> SB3 Contrib: https://github.com/Stable-Baselines-Team/stable-baselines3-contrib Install the RL Zoo (with SB3 and SB3-Contrib): ```bash pip install rl_zoo3 ``` ``` # Download model and save it into the logs/ folder python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga tatlook -f logs/ python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ ``` If you installed the RL Zoo3 via pip (`pip install rl_zoo3`), from anywhere you can do: ``` python -m rl_zoo3.load_from_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -orga tatlook -f logs/ python -m rl_zoo3.enjoy --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ ``` ## Training (with the RL Zoo) ``` python -m rl_zoo3.train --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ # Upload the model and generate video (when possible) python -m rl_zoo3.push_to_hub --algo dqn --env SpaceInvadersNoFrameskip-v4 -f logs/ -orga tatlook ``` ## Hyperparameters ```python OrderedDict([('batch_size', 32), ('buffer_size', 100000), ('env_wrapper', ['stable_baselines3.common.atari_wrappers.AtariWrapper']), ('exploration_final_eps', 0.01), ('exploration_fraction', 0.1), ('frame_stack', 4), ('gradient_steps', 1), ('learning_rate', 0.0001), ('learning_starts', 10000), ('n_timesteps', 1000000.0), ('optimize_memory_usage', False), ('policy', 'CnnPolicy'), ('target_update_interval', 1000), ('train_freq', 4), ('normalize', False)]) ``` # Environment Arguments ```python {'render_mode': 'rgb_array'} ```
{"library_name": "stable-baselines3", "tags": ["SpaceInvadersNoFrameskip-v4", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "DQN", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "SpaceInvadersNoFrameskip-v4", "type": "SpaceInvadersNoFrameskip-v4"}, "metrics": [{"type": "mean_reward", "value": "632.50 +/- 127.03", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
tatlook/dqn-SpaceInvadersNoFrameskip-v4
[ "stable-baselines3", "SpaceInvadersNoFrameskip-v4", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
2024-02-10T11:58:20+00:00
[]
[]
TAGS #stable-baselines3 #SpaceInvadersNoFrameskip-v4 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
# DQN Agent playing SpaceInvadersNoFrameskip-v4 This is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4 using the stable-baselines3 library and the RL Zoo. The RL Zoo is a training framework for Stable Baselines3 reinforcement learning agents, with hyperparameter optimization and pre-trained agents included. ## Usage (with SB3 RL Zoo) RL Zoo: URL SB3: URL SB3 Contrib: URL Install the RL Zoo (with SB3 and SB3-Contrib): If you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do: ## Training (with the RL Zoo) ## Hyperparameters # Environment Arguments
[ "# DQN Agent playing SpaceInvadersNoFrameskip-v4\nThis is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4\nusing the stable-baselines3 library\nand the RL Zoo.\n\nThe RL Zoo is a training framework for Stable Baselines3\nreinforcement learning agents,\nwith hyperparameter optimization and pre-trained agents included.", "## Usage (with SB3 RL Zoo)\n\nRL Zoo: URL\nSB3: URL\nSB3 Contrib: URL\n\nInstall the RL Zoo (with SB3 and SB3-Contrib):\n\n\n\n\nIf you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:", "## Training (with the RL Zoo)", "## Hyperparameters", "# Environment Arguments" ]
[ "TAGS\n#stable-baselines3 #SpaceInvadersNoFrameskip-v4 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n", "# DQN Agent playing SpaceInvadersNoFrameskip-v4\nThis is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4\nusing the stable-baselines3 library\nand the RL Zoo.\n\nThe RL Zoo is a training framework for Stable Baselines3\nreinforcement learning agents,\nwith hyperparameter optimization and pre-trained agents included.", "## Usage (with SB3 RL Zoo)\n\nRL Zoo: URL\nSB3: URL\nSB3 Contrib: URL\n\nInstall the RL Zoo (with SB3 and SB3-Contrib):\n\n\n\n\nIf you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:", "## Training (with the RL Zoo)", "## Hyperparameters", "# Environment Arguments" ]
[ 43, 90, 73, 9, 5, 7 ]
[ "passage: TAGS\n#stable-baselines3 #SpaceInvadersNoFrameskip-v4 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# DQN Agent playing SpaceInvadersNoFrameskip-v4\nThis is a trained model of a DQN agent playing SpaceInvadersNoFrameskip-v4\nusing the stable-baselines3 library\nand the RL Zoo.\n\nThe RL Zoo is a training framework for Stable Baselines3\nreinforcement learning agents,\nwith hyperparameter optimization and pre-trained agents included.## Usage (with SB3 RL Zoo)\n\nRL Zoo: URL\nSB3: URL\nSB3 Contrib: URL\n\nInstall the RL Zoo (with SB3 and SB3-Contrib):\n\n\n\n\nIf you installed the RL Zoo3 via pip ('pip install rl_zoo3'), from anywhere you can do:## Training (with the RL Zoo)## Hyperparameters# Environment Arguments" ]
[ 0.043572068214416504, 0.2414778620004654, -0.0026879787910729647, 0.012635791674256325, 0.05784223601222038, 0.0030472534708678722, 0.08585051447153091, 0.10650663822889328, 0.024212315678596497, -0.001382096204906702, 0.003954293206334114, 0.17533031105995178, 0.03632635250687599, 0.13125447928905487, -0.018073517829179764, -0.2066594809293747, -0.013479253277182579, -0.06247470900416374, -0.07153085619211197, 0.036099132150411606, 0.07206681370735168, -0.030116932466626167, 0.036061208695173264, -0.051406677812337875, -0.057161085307598114, 0.036824777722358704, -0.03157254680991173, 0.007067287806421518, 0.15158706903457642, -0.1222257912158966, 0.12329676002264023, 0.020955175161361694, 0.1896144151687622, -0.12332789599895477, 0.0339222252368927, 0.08982209116220474, -0.036988191306591034, 0.013221588917076588, 0.00975361280143261, -0.052562564611434937, 0.1590864509344101, -0.09371145814657211, 0.07146181166172028, 0.010926910676062107, -0.07592244446277618, -0.1774153709411621, -0.09356249868869781, 0.07947742193937302, 0.0617753230035305, 0.005319166928529739, 0.03726791962981224, 0.11306490749120712, -0.020991774275898933, 0.06488905102014542, 0.11562903225421906, -0.17549200356006622, 0.013578375801444054, 0.17859570682048798, 0.003242473118007183, 0.15767055749893188, -0.05546637624502182, 0.019877681508660316, 0.02752300351858139, 0.04758313298225403, 0.06873945891857147, -0.08186400681734085, -0.1364826112985611, -0.056155186146497726, -0.15456219017505646, -0.03352400287985802, 0.05195203423500061, -0.011860138736665249, -0.05783402919769287, -0.010724928230047226, -0.04010869935154915, 0.0008851495804265141, -0.028637725859880447, 0.01805497519671917, 0.07031578570604324, -0.01226285845041275, 0.02092539705336094, -0.08391954004764557, -0.0390290804207325, -0.038563769310712814, -0.018022390082478523, 0.12054917961359024, 0.08285853266716003, 0.0266572255641222, -0.04135355353355408, 0.10274127870798111, -0.07091585546731949, -0.05454207584261894, 0.04555258899927139, -0.03786851093173027, -0.10615779459476471, 0.02120024710893631, -0.05905991420149803, 0.026879185810685158, 0.09943640232086182, 0.18048083782196045, -0.09862488508224487, 0.012620617635548115, -0.03430783003568649, 0.08121664822101593, -0.03196052461862564, 0.03197542577981949, -0.0840383991599083, -0.016251085326075554, 0.17835216224193573, 0.0030782297253608704, 0.022272996604442596, 0.002074616262689233, -0.049819961190223694, -0.02881433069705963, -0.017756454646587372, 0.06631895154714584, 0.07032092660665512, 0.010587303899228573, -0.0037596761249005795, -0.027667716145515442, -0.036921944469213486, -0.05629328638315201, -0.04952820762991905, 0.018803736194968224, -0.04712437093257904, -0.047942135483026505, 0.06027210131287575, -0.005624116864055395, 0.11337806284427643, -0.025607796385884285, 0.026316547766327858, -0.019410157576203346, -0.07494441419839859, -0.13221681118011475, -0.0304415225982666, 0.0691632330417633, 0.04371757060289383, -0.22497159242630005, -0.16994807124137878, -0.008539012633264065, 0.017946386709809303, -0.018741264939308167, -0.11334165185689926, 0.02453240379691124, -0.007166135590523481, -0.049758363515138626, -0.01601579785346985, 0.10474669933319092, -0.020438622683286667, 0.018010856583714485, -0.05593825876712799, 0.16603368520736694, -0.14290283620357513, 0.031004127115011215, -0.08706212788820267, 0.023509707301855087, -0.21286657452583313, 0.041208744049072266, -0.177636057138443, 0.04863585904240608, -0.08500861376523972, 0.02327173389494419, 0.021320728585124016, 0.01968831568956375, 0.08580207824707031, 0.10143322497606277, -0.23631145060062408, 0.05405791476368904, 0.07900930196046829, -0.022739801555871964, -0.04218491166830063, 0.06798892468214035, -0.06558530032634735, 0.1382148116827011, 0.046505436301231384, 0.24831900000572205, 0.10361487418413162, -0.2036508023738861, 0.061786454170942307, 0.0578593946993351, -0.08880111575126648, -0.004730981774628162, -0.020022382959723473, 0.11598580330610275, -0.01114928349852562, 0.03338807821273804, -0.12186288088560104, 0.1456439197063446, 0.02738998830318451, -0.0165485180914402, -0.04454165697097778, -0.1614885926246643, 0.10309953987598419, -0.015504824928939342, 0.09532155096530914, -0.042415786534547806, 0.0001161050095106475, -0.011168917641043663, 0.18012429773807526, -0.043841805309057236, 0.0007168867159634829, 0.07871408760547638, 0.10895700752735138, 0.028009075671434402, -0.020230965688824654, -0.20380273461341858, -0.0423048660159111, 0.02367858961224556, 0.044489551335573196, 0.2190362960100174, 0.19936694204807281, 0.07770156860351562, -0.022313760593533516, -0.025487221777439117, -0.003248062450438738, -0.05106664076447487, 0.03467361256480217, -0.027858436107635498, -0.024532482028007507, 0.06065356358885765, -0.09305168688297272, 0.02817818708717823, -0.13112716376781464, 0.06307920068502426, -0.17345242202281952, 0.06863926351070404, 0.021998396143317223, -0.005436043255031109, 0.024577690288424492, -0.011292695067822933, -0.034188106656074524, -0.06233125180006027, 0.07110602408647537, 0.06098933145403862, 0.014702376909554005, 0.0021991983521729708, -0.0683600977063179, -0.13828523457050323, 0.08231553435325623, -0.04042381793260574, -0.14305958151817322, 0.06392676383256912, 0.011172642931342125, 0.04875864461064339, -0.05975872278213501, 0.016254881396889687, 0.22900153696537018, 0.05321883037686348, 0.09785865992307663, -0.04092191904783249, -0.022525805979967117, -0.06617844104766846, -0.06677833944559097, 0.09694591909646988, 0.10812206566333771, 0.060318704694509506, -0.0030071530491113663, 0.07626225054264069, 0.10942911356687546, -0.1035122498869896, -0.0651884600520134, 0.03220061957836151, -0.05973697826266289, 0.019652515649795532, 0.049140311777591705, 0.02971293032169342, 0.08619047701358795, 0.1833551675081253, 0.008245792239904404, 0.0386311337351799, -0.025997694581747055, 0.026109617203474045, -0.15547916293144226, -0.03145433962345123, 0.04308181628584862, 0.00886955764144659, -0.07408110797405243, 0.04994636029005051, 0.051439400762319565, 0.13607151806354523, -0.08217083662748337, -0.13170577585697174, -0.059745315462350845, -0.03804200142621994, -0.04239124804735184, 0.14975430071353912, -0.08507520705461502, -0.19221234321594238, -0.017164425924420357, -0.15751953423023224, -0.02518727444112301, -0.005179801490157843, 0.002318724524229765, -0.08325926214456558, 0.017780914902687073, 0.010001576505601406, -0.03129372000694275, -0.0684933215379715, -0.06596160680055618, -0.05786636844277382, 0.09124112874269485, 0.06932931393384933, -0.12240120023488998, -0.00961651187390089, -0.03742414712905884, -0.020465577021241188, 0.04516167193651199, 0.08452648669481277, -0.007267598994076252, 0.07773483544588089, -0.13209199905395508, -0.06962883472442627, 0.02834828943014145, 0.2766247093677521, 0.02882981114089489, 0.004668009467422962, 0.17051753401756287, -0.03629542142152786, 0.04912714660167694, 0.16181479394435883, 0.030781643465161324, -0.14196757972240448, 0.07090470939874649, -0.011341600678861141, -0.09542687982320786, -0.1706860214471817, -0.10215658694505692, -0.037867411971092224, -0.05015881359577179, 0.05638284236192703, 0.004951419774442911, -0.04476970434188843, 0.05910305306315422, 0.08782228082418442, -0.017004497349262238, -0.06151578947901726, 0.11129767447710037, 0.032263003289699554, -0.030136963352560997, 0.08078382909297943, -0.042354047298431396, -0.04206389561295509, 0.0032403599470853806, 0.22643887996673584, 0.0937788337469101, -0.01775507442653179, -0.042567066848278046, 0.019317636266350746, 0.05095715448260307, 0.03613382205367088, 0.11312435567378998, -0.06975842267274857, -0.06826137751340866, -0.035185977816581726, 0.027829548344016075, -0.02945687249302864, 0.08205190300941467, 0.0630207508802414, 0.005563626065850258, -0.04653681069612503, -0.07972332090139389, -0.04849022626876831, 0.08408913016319275, -0.027642227709293365, -0.10093270242214203, 0.09321888536214828, 0.048575710505247116, 0.0016974330646917224, 0.03055831417441368, 0.027994604781270027, 0.01462269201874733, -0.07982148975133896, -0.06775744259357452, 0.011468625627458096, 0.07076629996299744, -0.06822766363620758, -0.027886953204870224, -0.19817815721035004, 0.14578363299369812, 0.010630400851368904, 0.04118429124355316, -0.13048617541790009, 0.1209396943449974, -0.023116756230592728, -0.026430301368236542, 0.013811616227030754, 0.0014643745962530375, 0.08203291147947311, -0.04806509613990784, 0.15762180089950562, 0.009528410620987415, -0.28092408180236816, -0.1418946087360382, -0.08416824042797089, -0.051183976233005524, -0.022873088717460632, 0.014752174727618694, 0.0642135739326477, 0.01516205258667469, 0.003868846921250224, -0.013076163828372955, 0.03185269236564636, -0.09826882928609848, -0.06493937969207764, -0.04839126765727997, -0.02250157669186592, -0.06525848805904388, -0.05647949501872063, -0.0006809153710491955, -0.17226077616214752, 0.12522587180137634, 0.11787347495555878, -0.06451737880706787, -0.041814323514699936, -0.06554657220840454, 0.046191465109586716, -0.07571537792682648, 0.0469326451420784, 0.003414976177737117, 0.019198855385184288, -0.06806991249322891, -0.17922484874725342, 0.016097763553261757, -0.10899919271469116, 0.03772687539458275, -0.05070559307932854, 0.020257100462913513, 0.08594245463609695, 0.17520126700401306, 0.05856714025139809, 0.01460097823292017, -0.07239776104688644, -0.07543374598026276, -0.0017121878918260336, -0.06344114243984222, 0.05762333422899246, -0.009151889942586422, -0.20333483815193176, 0.02763226442039013, -0.11414948850870132, 0.06860900670289993, 0.3310066759586334, 0.3324824273586273, -0.10698744654655457, 0.1177443116903305, 0.04819539934396744, -0.042202454060316086, -0.21051374077796936, -0.002244179602712393, 0.012272895313799381, 0.024992236867547035, 0.13725964725017548, -0.12924811244010925, 0.05453680083155632, 0.0794181227684021, -0.024458877742290497, 0.01456840243190527, -0.09078162908554077, -0.10816970467567444, 0.20847418904304504, 0.14226987957954407, 0.04421741142868996, -0.09421348571777344, 0.08391669392585754, 0.004295284394174814, 0.08375877887010574, 0.2107764035463333, -0.052112679928541183, 0.10695768147706985, 0.005195184610784054, 0.19852910935878754, 0.0328996516764164, -0.023768596351146698, 0.10834760218858719, -0.009801650419831276, 0.07911337912082672, 0.03985166177153587, -0.007676942739635706, 0.010487722232937813, -0.04522453248500824, 0.014148596674203873, -0.028376007452607155, 0.010284217074513435, -0.2274095118045807, 0.0582297146320343, -0.06368855386972427, 0.04604509472846985, 0.008256820961833, -0.0999874547123909, -0.03583388403058052, 0.06431841105222702, 0.08014573156833649, 0.01975327916443348, 0.0436067171394825, -0.03867863491177559, 0.11051398515701294, 0.20660489797592163, -0.009811338968575, 0.17751595377922058, -0.0615963339805603, 0.01464168168604374, -0.023011628538370132, -0.04223164543509483, -0.1462583988904953, -0.035259708762168884, 0.03498423472046852, 0.057734888046979904, 0.015203364193439484, 0.049647457897663116, -0.05656236410140991, 0.08498423546552658, 0.021687336266040802, -0.041541360318660736, 0.033579520881175995, 0.08835696429014206, 0.12415177375078201, 0.010754258371889591, -0.030121933668851852, 0.06147436052560806, -0.08128108084201813, -0.09446098655462265, -0.004497923422604799, -0.029991207644343376, -0.1083834245800972, 0.11353230476379395, 0.16914646327495575, 0.039594944566488266, -0.057076629251241684, 0.10688766092061996, -0.02768099494278431, 0.10047874599695206, 0.009198128245770931, 0.06507332623004913, -0.014091075398027897, -0.03691792115569115, 0.10611724853515625, -0.05442855879664421, -0.01637818105518818, 0.07645545154809952, -0.06522727757692337, -0.023877469822764397, -0.0801999643445015, 0.06034626066684723, 0.09222240000963211, -0.16854619979858398, -0.0639432892203331, -0.032122284173965454, -0.08628080040216446, 0.013965039514005184, 0.012447911314666271, 0.0710059329867363, -0.08589600026607513, 0.06316167116165161, -0.024337708950042725, 0.015639442950487137, -0.03689891844987869, 0.019222697243094444, -0.19525384902954102, -0.002140450058504939, -0.11280795186758041, -0.00348020251840353, -0.002931603929027915, 0.04463808611035347, -0.04961875081062317, -0.029358822852373123, -0.0030675032176077366, 0.044366419315338135, -0.16609135270118713, 0.002798673929646611, -0.011639905162155628, 0.03210212290287018, -0.0002893915225286037, -0.0983390137553215, 0.014195028692483902, -0.04294256120920181, -0.04198618605732918, 0.04925514757633209, 0.009436776861548424, 0.06470516324043274, -0.2795179784297943, -0.14905457198619843, 0.030816160142421722, 0.0683867484331131, 0.05483196675777435, -0.1830425262451172, 0.03568267077207565, -0.08042316138744354, -0.02253127470612526, -0.037770628929138184, 0.018491698428988457, -0.0539514496922493, 0.0018174031283706427, -0.04225044324994087, -0.023033907637000084, -0.028055014088749886, -0.07556360960006714, 0.0826747715473175, 0.12462522834539413, 0.07555580884218216, -0.03807181864976883, 0.09595896303653717, -0.10009756684303284, -0.04657831788063049, -0.04052736237645149, -0.036951083689928055, 0.017965637147426605, -0.0870552659034729, 0.048530060797929764, 0.05188591405749321, 0.18719671666622162, -0.08520494401454926, -0.058800119906663895, -0.014255574904382229, 0.0746525228023529, 0.07849094271659851, 0.005095830652862787, 0.17779210209846497, -0.045693784952163696, 0.05693846940994263, 0.021304311230778694, 0.046699028462171555, 0.10497613251209259, -0.023569339886307716, 0.14490213990211487, 0.21171095967292786, -0.037196725606918335, -0.11048602312803268, 0.043668005615472794, 0.01745123788714409, -0.002401199424639344, 0.05968761444091797, 0.11983796209096909, -0.050589341670274734, -0.10903856158256531, 0.23442286252975464, 0.054169271141290665, -0.11218088120222092, 0.09546315670013428, 0.039532262831926346, -0.015890996903181076, -0.1301896870136261, 0.010444961488246918, -0.0013640925753861666, -0.11233190447092056, 0.03386834263801575, -0.06087532266974449, -0.025547027587890625, 0.11809267848730087, 0.008789865300059319, 0.03317064419388771, -0.04139537364244461, -0.03756232187151909, -0.04352104663848877, -0.04273213446140289, -0.012549578212201595, -0.02991986647248268, -0.030186517164111137, -0.07621737569570541, -0.007770835887640715, -0.012012424878776073, 0.030795488506555557, -0.015285328030586243, -0.02503054589033127, -0.021192016080021858, -0.06697061657905579, -0.0026312144473195076, -0.008178025484085083, 0.015549594536423683, 0.010121971368789673, 0.2358063906431198, 0.07042546570301056, -0.10260069370269775, -0.01036880537867546, 0.22197756171226501, -0.03853277862071991, -0.06528383493423462, -0.07849395275115967, 0.25128230452537537, -0.10482002794742584, 0.051095426082611084, -0.005819917656481266, -0.06550488620996475, -0.07153836637735367, 0.2309868484735489, 0.13502730429172516, -0.1677926480770111, 0.06329060345888138, -0.0368385910987854, -0.009490780532360077, -0.14286863803863525, 0.16013580560684204, 0.1865294873714447, 0.09480160474777222, -0.12259847670793533, 0.0023130534682422876, -0.03518044203519821, -0.018328361213207245, -0.1660851687192917, -0.004593863617628813, -0.029364850372076035, -0.0427238829433918, -0.050771355628967285, 0.029773715883493423, -0.15205919742584229, -0.0927426889538765, -0.1916799396276474, -0.11482496559619904, -0.12386849522590637, -0.04549141973257065, -0.11142764985561371, -0.0019938007462769747, 0.02257080189883709, -0.0641874223947525, 0.021061956882476807, -0.0212461706250906, -0.05887424945831299, 0.015386379323899746, -0.08395619690418243, 0.0674985870718956, 0.06488548219203949, 0.15327942371368408, -0.0790991559624672, 0.025424562394618988, 0.07090727984905243, -0.057595450431108475, -0.10164349526166916, 0.06067253649234772, 0.015708057209849358, -0.1972588747739792, 0.007548294495791197, 0.17712996900081635, -0.10420889407396317, 0.09745754301548004, 0.048501528799533844, -0.012951982207596302, 0.0867827981710434, -0.024721821770071983, -0.016682926565408707, -0.04852180927991867, -0.011212974786758423, -0.10143939405679703, 0.09892100840806961, 0.0876845121383667, -0.0517118014395237, 0.07436849176883698, -0.09508965909481049, -0.04068392515182495, 0.13103286921977997, -0.010057874955236912, -0.08450483530759811, -0.11667824536561966, -0.04081142693758011, 0.09684515744447708, -0.018041390925645828, -0.20185889303684235, -0.11639472097158432, -0.11752668023109436, -0.00014377340266946703, -0.03563340753316879, 0.061800602823495865, 0.02430674433708191, -0.02556120604276657, -0.008150683715939522, -0.17615078389644623, -0.06614746153354645, 0.13479791581630707, -0.10176112502813339, -0.07456064969301224 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text-generation
aidonuts/marigold-005-ep3
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T11:58:24+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 60, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.04654794931411743, 0.16618601977825165, -0.005445904564112425, 0.01853804849088192, 0.0981811136007309, 0.011998992413282394, 0.06433123350143433, 0.11398410052061081, -0.0230073444545269, 0.11406639218330383, 0.03047988750040531, 0.10172267258167267, 0.11317981779575348, 0.14841650426387787, -0.002152352826669812, -0.22403094172477722, 0.050844956189394, -0.12105348706245422, -0.033293843269348145, 0.11749980598688126, 0.1483822613954544, -0.09928343445062637, 0.07274559140205383, -0.029687678441405296, -0.012143402360379696, -0.030057786032557487, -0.05890674889087677, -0.046214159578084946, 0.04651786759495735, 0.06640566885471344, 0.06770290434360504, 0.0071083661168813705, 0.09012923389673233, -0.2696533799171448, 0.018959321081638336, 0.07145345956087112, -0.002759667346253991, 0.06957992166280746, 0.06404146552085876, -0.07107418030500412, 0.10337356477975845, -0.05106033384799957, 0.14650006592273712, 0.08365883678197861, -0.09081148356199265, -0.1895141303539276, -0.08866965025663376, 0.09882009029388428, 0.17572562396526337, 0.04925641790032387, -0.02320658043026924, 0.09761467576026917, -0.08769196271896362, 0.015438909642398357, 0.04981724172830582, -0.07620415836572647, -0.05378096550703049, 0.05986575037240982, 0.07907199114561081, 0.06627275794744492, -0.12434766441583633, -0.02885502204298973, 0.005009706597775221, 0.010980482213199139, 0.0769270583987236, 0.01728810742497444, 0.146672785282135, 0.0338633768260479, -0.12615777552127838, -0.04880760237574577, 0.09869225323200226, 0.03395522013306618, -0.04422314465045929, -0.24749068915843964, -0.03152675926685333, -0.030810698866844177, -0.029386121779680252, -0.03716538846492767, 0.04340358078479767, -0.007673026993870735, 0.08638741075992584, -0.0060646249912679195, -0.07403432577848434, -0.03937075287103653, 0.06169692054390907, 0.0672287791967392, 0.02999979443848133, -0.013745363801717758, 0.010938193649053574, 0.11620724946260452, 0.1095694974064827, -0.12054188549518585, -0.05555335059762001, -0.06393084675073624, -0.08656639605760574, -0.040790557861328125, 0.034162238240242004, 0.03456587344408035, 0.05349370837211609, 0.25305667519569397, 0.015654386952519417, 0.059652652591466904, 0.034477248787879944, 0.007892133668065071, 0.05848940089344978, 0.11044429242610931, -0.06018859148025513, -0.10444226115942001, -0.02648012898862362, 0.08843598514795303, 0.008199662901461124, -0.03287925571203232, -0.05088530853390694, 0.06019928678870201, 0.01946467161178589, 0.11926145106554031, 0.09061790257692337, 0.010536285117268562, -0.07121123373508453, -0.061038948595523834, 0.1891259253025055, -0.16544590890407562, 0.04322727024555206, 0.035097137093544006, -0.03903156518936157, 0.00019933005387429148, 0.013914269395172596, 0.016625655815005302, -0.025983380153775215, 0.09017423540353775, -0.054113563150167465, -0.04145489260554314, -0.11186197400093079, -0.03383193537592888, 0.033762916922569275, 0.008953776210546494, -0.035059962421655655, -0.033713940531015396, -0.08351044356822968, -0.07577689737081528, 0.09320491552352905, -0.07346344739198685, -0.04878907650709152, -0.01804324984550476, -0.07530532777309418, 0.022395428270101547, 0.019394835457205772, 0.07707412540912628, -0.02362251654267311, 0.04399976506829262, -0.05189276114106178, 0.05863580107688904, 0.11207318305969238, 0.03570080175995827, -0.05736649036407471, 0.06062258034944534, -0.23834340274333954, 0.09552820026874542, -0.07409077137708664, 0.05591456592082977, -0.153293639421463, -0.024439791217446327, 0.04788333550095558, 0.008784620091319084, -0.009650949388742447, 0.13416339457035065, -0.21702027320861816, -0.02536402828991413, 0.1717337965965271, -0.10057014971971512, -0.07069246470928192, 0.05619903281331062, -0.04835370555520058, 0.10988964140415192, 0.03825836628675461, -0.025690359994769096, 0.06171267107129097, -0.1267417073249817, 0.003717758459970355, -0.05005312338471413, -0.017048977315425873, 0.1548657864332199, 0.07182947546243668, -0.07217690348625183, 0.07399354875087738, 0.025708531960844994, -0.0246540866792202, -0.04625825211405754, -0.015164627693593502, -0.10536660254001617, 0.014689887873828411, -0.06369215250015259, 0.014470234513282776, -0.020807426422834396, -0.09071163833141327, -0.027962757274508476, -0.17504668235778809, -0.03014434315264225, 0.08651752024888992, -0.008693269453942776, -0.01803150773048401, -0.1178668737411499, 0.009341353550553322, 0.04177580401301384, 0.0061247628182172775, -0.13462838530540466, -0.04812471568584442, 0.02780051715672016, -0.1600649207830429, 0.034652888774871826, -0.05392369255423546, 0.04932025074958801, 0.025790516287088394, -0.028889117762446404, -0.026493212208151817, 0.021633783355355263, 0.005992184858769178, -0.011999987065792084, -0.24343903362751007, -0.028118690475821495, -0.024888472631573677, 0.1682123839855194, -0.20917098224163055, 0.03546025976538658, 0.07867541164159775, 0.15366052091121674, 0.011240328662097454, -0.04177491366863251, 0.005974748637527227, -0.06935794651508331, -0.02736494317650795, -0.05875484645366669, -0.0047869328409433365, -0.03310677409172058, -0.04545191675424576, 0.04568447172641754, -0.16510973870754242, -0.032636504620313644, 0.09776268899440765, 0.06289951503276825, -0.13922683894634247, -0.020621931180357933, -0.03630133345723152, -0.049253206700086594, -0.04911839962005615, -0.0605199858546257, 0.10893940925598145, 0.05891856551170349, 0.04574795812368393, -0.05928509309887886, -0.07568105310201645, -0.001827909960411489, -0.013898161239922047, -0.017864689230918884, 0.09759635478258133, 0.0751434788107872, -0.13251115381717682, 0.09224759042263031, 0.09603385627269745, 0.07919023185968399, 0.09113933145999908, -0.02355697751045227, -0.08261934667825699, -0.045987509191036224, 0.031442027539014816, 0.020124373957514763, 0.13039541244506836, -0.024294709786772728, 0.04352088272571564, 0.042134687304496765, -0.019369594752788544, 0.014752166345715523, -0.08687400817871094, 0.033972494304180145, 0.028472330421209335, -0.016721390187740326, 0.050190530717372894, -0.03876714035868645, 0.02440318465232849, 0.08830609917640686, 0.045322712510824203, 0.03507532551884651, 0.015493292361497879, -0.05206458270549774, -0.1083620935678482, 0.16405931115150452, -0.12714070081710815, -0.22483378648757935, -0.13936103880405426, 0.0037376401014626026, 0.035628627985715866, -0.015835661441087723, 0.002417160663753748, -0.059374887496232986, -0.12220635265111923, -0.08858037739992142, 0.015140829607844353, 0.04942670464515686, -0.09028962254524231, -0.06437795609235764, 0.058117836713790894, 0.03889724239706993, -0.14560972154140472, 0.017612040042877197, 0.04854894429445267, -0.09789852797985077, -0.006774199660867453, 0.08094939589500427, 0.0698540136218071, 0.1770169734954834, 0.017703235149383545, -0.021850809454917908, 0.032354529947042465, 0.20614571869373322, -0.13538233935832977, 0.11083246022462845, 0.13607586920261383, -0.09041404724121094, 0.08072979003190994, 0.19951270520687103, 0.03932560607790947, -0.10153959691524506, 0.031980328261852264, 0.02283124253153801, -0.0284719280898571, -0.24526868760585785, -0.07212468236684799, -0.004402178805321455, -0.058010730892419815, 0.07660572230815887, 0.09286724030971527, 0.08215958625078201, 0.012304253876209259, -0.09310996532440186, -0.08154371380805969, 0.05942574888467789, 0.10367169976234436, 0.024584239348769188, -0.010839897207915783, 0.08998730033636093, -0.034100502729415894, 0.019626356661319733, 0.0853661298751831, 0.005239574704319239, 0.17840281128883362, 0.05159219726920128, 0.18830420076847076, 0.07925192266702652, 0.07219027727842331, 0.009912233799695969, 0.013080619275569916, 0.018877580761909485, 0.03300119563937187, -0.002769160782918334, -0.08440786600112915, -0.02248465269804001, 0.11566436290740967, 0.06668911874294281, 0.010815348476171494, 0.015172341838479042, -0.04104290530085564, 0.07965951412916183, 0.1831512451171875, -0.007656289264559746, -0.1783534437417984, -0.057547420263290405, 0.07553383708000183, -0.09879875183105469, -0.09854305535554886, -0.013454320840537548, 0.03072015568614006, -0.17046253383159637, 0.023390959948301315, -0.02239842526614666, 0.1106182336807251, -0.14194999635219574, -0.020490378141403198, 0.07218493521213531, 0.07199500501155853, 0.004729843698441982, 0.05758659541606903, -0.16417601704597473, 0.10671813786029816, 0.008950476534664631, 0.06779605895280838, -0.09610627591609955, 0.1008887067437172, -0.004196076653897762, -0.02063460275530815, 0.1393408179283142, 0.002700034761801362, -0.06884108483791351, -0.0763031542301178, -0.08754398673772812, -0.009632662869989872, 0.12754282355308533, -0.1419651061296463, 0.08767123520374298, -0.037212442606687546, -0.0424150750041008, -0.0017086371080949903, -0.10206665843725204, -0.11638247221708298, -0.18888559937477112, 0.06001543253660202, -0.13492922484874725, 0.03152317553758621, -0.10799519717693329, -0.032371897250413895, -0.030304040759801865, 0.19337286055088043, -0.23447458446025848, -0.07199826091527939, -0.1475764364004135, -0.10233612358570099, 0.1443224400281906, -0.0501345656812191, 0.08485390990972519, -0.007241467013955116, 0.16846685111522675, 0.019060896709561348, -0.02531743235886097, 0.0971490666270256, -0.09173708409070969, -0.19302815198898315, -0.07869284600019455, 0.15662524104118347, 0.13260218501091003, 0.031680017709732056, -0.002461588243022561, 0.036563750356435776, -0.015421539545059204, -0.11935004591941833, 0.015969349071383476, 0.1787186712026596, 0.06237189099192619, 0.02331034652888775, -0.027346095070242882, -0.11273157596588135, -0.06900003552436829, -0.028530338779091835, 0.03054865077137947, 0.17762407660484314, -0.07057618349790573, 0.18207968771457672, 0.14163152873516083, -0.05922834202647209, -0.20400173962116241, 0.010538800619542599, 0.03055560030043125, 0.0009220078936778009, 0.02591954916715622, -0.20123432576656342, 0.08688826113939285, 0.004683020059019327, -0.05110127478837967, 0.13194532692432404, -0.17217805981636047, -0.14451217651367188, 0.0765485092997551, 0.038384392857551575, -0.19559739530086517, -0.12913893163204193, -0.09174312651157379, -0.045869920402765274, -0.18591414391994476, 0.09569250047206879, 0.0305706188082695, 0.010893458500504494, 0.03030681423842907, 0.029179483652114868, 0.019487828016281128, -0.0418255440890789, 0.18391458690166473, -0.024792250245809555, 0.026594700291752815, -0.08539514988660812, -0.06927408277988434, 0.03743394836783409, -0.052842434495687485, 0.07349982857704163, -0.023486759513616562, 0.007861839607357979, -0.10348054021596909, -0.042148489505052567, -0.03735732287168503, 0.015448716469109058, -0.09657872468233109, -0.08514349907636642, -0.045032672584056854, 0.09675803780555725, 0.09690850973129272, -0.033646680414676666, -0.028050623834133148, -0.07533035427331924, 0.04412057250738144, 0.19926515221595764, 0.1785389482975006, 0.042153384536504745, -0.08034496754407883, -0.004150947090238333, -0.010121207684278488, 0.04310847446322441, -0.20463712513446808, 0.06283636391162872, 0.05450061708688736, 0.01973269321024418, 0.11436162889003754, -0.019565396010875702, -0.15359151363372803, -0.07263088971376419, 0.06303015351295471, -0.060181066393852234, -0.19620554149150848, 0.00867035984992981, 0.060603946447372437, -0.16371412575244904, -0.04535605385899544, 0.04643881320953369, -0.005620351992547512, -0.038163937628269196, 0.021896906197071075, 0.09194854646921158, 0.0026654244866222143, 0.07427921891212463, 0.05387866869568825, 0.0827430784702301, -0.10537070035934448, 0.08090532571077347, 0.08839722722768784, -0.08452684432268143, 0.023530138656497, 0.10478579998016357, -0.059433579444885254, -0.03440561518073082, 0.020135708153247833, 0.08153781294822693, 0.01775863952934742, -0.040019966661930084, 0.013229827396571636, -0.10452935844659805, 0.05954122915863991, 0.08839859813451767, 0.032507482916116714, 0.016702456399798393, 0.03425082191824913, 0.04607953503727913, -0.07238735258579254, 0.12142276018857956, 0.031868141144514084, 0.017129309475421906, -0.036505792289972305, -0.040896978229284286, 0.019542274996638298, -0.03214648738503456, -0.005015232600271702, -0.03023446537554264, -0.07695909589529037, -0.014793801121413708, -0.1626158058643341, -0.011131818406283855, -0.05648450180888176, 0.010329355485737324, 0.03204665705561638, -0.032609567046165466, 0.008124498650431633, 0.009250079281628132, -0.07695289701223373, -0.0663459524512291, -0.020460480824112892, 0.09540658444166183, -0.16213038563728333, 0.022481130436062813, 0.08244425803422928, -0.12187694013118744, 0.09281346201896667, 0.016204802319407463, -0.006236857734620571, 0.025038830935955048, -0.1475188434123993, 0.034843120723962784, -0.03386561945080757, 0.010836300440132618, 0.04373383894562721, -0.21569781005382538, -0.00004886732858722098, -0.033673107624053955, -0.06639216095209122, -0.009451326914131641, -0.03672455996274948, -0.11508306115865707, 0.1058407872915268, 0.007236586883664131, -0.08753558248281479, -0.03186136856675148, 0.029325377196073532, 0.0838974118232727, -0.021959776058793068, 0.15145497024059296, -0.008370938710868359, 0.07429654151201248, -0.16209737956523895, -0.018623165786266327, -0.006028574425727129, 0.022658247500658035, -0.01664556935429573, -0.01111356820911169, 0.044031109660863876, -0.022746501490473747, 0.17925859987735748, -0.030318550765514374, 0.02272745408117771, 0.06815794110298157, 0.019072026014328003, -0.030184008181095123, 0.10406795144081116, 0.04094860330224037, 0.02014910988509655, 0.018591465428471565, 0.003289656015112996, -0.04647882282733917, -0.03173251822590828, -0.19407226145267487, 0.07288651913404465, 0.15608493983745575, 0.09729263186454773, -0.016707008704543114, 0.07954329252243042, -0.10199416428804398, -0.1109243705868721, 0.12477338314056396, -0.04797708988189697, -0.002418199321255088, -0.07150927931070328, 0.13247236609458923, 0.1437523066997528, -0.1859612911939621, 0.07269313186407089, -0.0699717253446579, -0.04708027467131615, -0.10980689525604248, -0.19441905617713928, -0.05561789125204086, -0.049456022679805756, -0.016053348779678345, -0.04698808491230011, 0.07504211366176605, 0.054538097232580185, 0.006766852922737598, -0.0023397188633680344, 0.06506035476922989, -0.031050674617290497, -0.0037882844917476177, 0.032597362995147705, 0.06591679900884628, 0.012734474614262581, -0.030802709981799126, 0.016619903966784477, -0.013545602560043335, 0.045626189559698105, 0.06578011065721512, 0.04976864159107208, -0.02938537672162056, 0.014603170566260815, -0.038539156317710876, -0.10249634087085724, 0.043612558394670486, -0.024421939626336098, -0.0789753645658493, 0.15477414429187775, 0.023680059239268303, 0.007779473438858986, -0.020137663930654526, 0.23901568353176117, -0.0738423764705658, -0.0964353010058403, -0.14737580716609955, 0.10557299107313156, -0.038081806153059006, 0.05800395458936691, 0.04625935107469559, -0.10226529091596603, 0.018044332042336464, 0.1338089406490326, 0.16182038187980652, -0.039008259773254395, 0.020095856860280037, 0.031135575845837593, 0.00566398398950696, -0.03622615709900856, 0.04847532883286476, 0.06906453520059586, 0.16569648683071136, -0.04632584750652313, 0.09100406616926193, 0.0019041687482967973, -0.09579581767320633, -0.038361791521310806, 0.11069868505001068, -0.016052277758717537, 0.019335128366947174, -0.05818064883351326, 0.11742528527975082, -0.06386786699295044, -0.23783175647258759, 0.06453443318605423, -0.0684293657541275, -0.13765870034694672, -0.02378307841718197, 0.08207765966653824, -0.012955902144312859, 0.027587108314037323, 0.0730307325720787, -0.07240920513868332, 0.201939657330513, 0.03798431158065796, -0.05499868467450142, -0.055047210305929184, 0.0805421993136406, -0.10008571296930313, 0.2739645540714264, 0.01557221356779337, 0.04601577669382095, 0.10384146869182587, -0.009341772645711899, -0.13838784396648407, 0.019836371764540672, 0.09581108391284943, -0.10502193123102188, 0.04196618124842644, 0.19815568625926971, -0.0014755994779989123, 0.12389086186885834, 0.07657600939273834, -0.07551808655261993, 0.0478031262755394, -0.08054235577583313, -0.06760486960411072, -0.09260394424200058, 0.09703279286623001, -0.07772123068571091, 0.14251399040222168, 0.13876807689666748, -0.05074559152126312, 0.012724342755973339, -0.031311117112636566, 0.044293127954006195, -0.00010600237874314189, 0.10321761667728424, 0.004272161517292261, -0.1832672357559204, 0.024692710489034653, 0.005650998093187809, 0.10749758034944534, -0.16033467650413513, -0.09566054493188858, 0.042343202978372574, 0.003505636239424348, -0.0672195628285408, 0.1290110945701599, 0.05665452033281326, 0.04342988133430481, -0.03997718170285225, -0.03521440550684929, -0.0060732318088412285, 0.13561366498470306, -0.10713256150484085, 0.0009933578548952937 ]
null
null
null
# **Q-Learning** Agent playing1 **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="tahsinkheya/autonomous-taxi", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
{"tags": ["Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "autonomous-taxi", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "Taxi-v3", "type": "Taxi-v3"}, "metrics": [{"type": "mean_reward", "value": "7.52 +/- 2.67", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
tahsinkheya/autonomous-taxi
[ "Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
2024-02-10T12:00:11+00:00
[]
[]
TAGS #Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
# Q-Learning Agent playing1 Taxi-v3 This is a trained model of a Q-Learning agent playing Taxi-v3 . ## Usage
[ "# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage" ]
[ "TAGS\n#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n", "# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage" ]
[ 32, 33 ]
[ "passage: TAGS\n#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage" ]
[ 0.048862796276807785, -0.16549694538116455, -0.005485367961227894, 0.02960980497300625, 0.1345081776380539, -0.01784728653728962, 0.11895976960659027, 0.07759871333837509, -0.07461097836494446, -0.055395450443029404, 0.1418241262435913, 0.09088201075792313, 0.055222880095243454, 0.05699880048632622, 0.09511256217956543, -0.27440664172172546, 0.048217080533504486, -0.02918700873851776, 0.05621987581253052, 0.11878681182861328, 0.0670095682144165, -0.040441032499074936, 0.061956584453582764, 0.11818158626556396, -0.1018151044845581, -0.007344264071434736, 0.035402704030275345, -0.09440053254365921, 0.17413531243801117, 0.07204403728246689, 0.12337774783372879, 0.05132639780640602, 0.179361954331398, -0.12762396037578583, 0.024310702458024025, -0.0010275895474478602, -0.10138072073459625, -0.03909514099359512, -0.012415820732712746, -0.08349097520112991, 0.03230205550789833, 0.23522862792015076, 0.07199250161647797, 0.06632792949676514, -0.17707863450050354, -0.06584878265857697, -0.04375573247671127, 0.069611094892025, 0.14951466023921967, 0.03758616745471954, -0.033800311386585236, 0.1684885323047638, -0.2564343810081482, 0.05066783353686333, 0.037275806069374084, -0.42313119769096375, 0.017119819298386574, 0.1507398933172226, 0.15090937912464142, 0.06909667700529099, -0.10573802888393402, 0.013512322679162025, 0.051325585693120956, -0.0005318621988408267, 0.024325110018253326, 0.006554204970598221, 0.15601307153701782, 0.08537693321704865, -0.1487821787595749, -0.058576688170433044, 0.17441977560520172, -0.03788546845316887, -0.02613203600049019, -0.039745692163705826, 0.0067160045728087425, -0.06427708268165588, -0.004067842848598957, -0.1777995079755783, 0.00734262028709054, 0.06666424125432968, -0.014348524622619152, 0.014901017770171165, -0.035522811114788055, -0.0966939702630043, -0.023098144680261612, -0.08592145889997482, 0.01677769608795643, -0.006319406442344189, -0.10187895596027374, 0.05002119392156601, -0.061138734221458435, 0.0014382408699020743, -0.05123179033398628, -0.15047866106033325, -0.049055423587560654, -0.03481535613536835, 0.1474713832139969, -0.0044205985032022, -0.01873963139951229, -0.03164304047822952, 0.15474793314933777, 0.049551334232091904, -0.05370146036148071, 0.05625450983643532, 0.07605006545782089, 0.23867930471897125, 0.10401605814695358, 0.10196955502033234, -0.06798075139522552, 0.10180158913135529, -0.12330973148345947, -0.08915644884109497, -0.17508824169635773, 0.11820860952138901, 0.00015364694991149008, 0.1317785084247589, -0.12023144960403442, 0.07898581773042679, -0.067511186003685, 0.013453764840960503, 0.01636839471757412, 0.0820009782910347, -0.012399360537528992, 0.10676060616970062, -0.005061192903667688, -0.06941985338926315, 0.014177112840116024, 0.05935845896601677, 0.03754841163754463, -0.038601722568273544, -0.03192409873008728, -0.05762290954589844, -0.05065649375319481, -0.10128600150346756, -0.06447898596525192, 0.018573462963104248, -0.007677143905311823, -0.1833900660276413, -0.06407523155212402, 0.00897200871258974, 0.015712225809693336, -0.03988850116729736, -0.05148044601082802, -0.15265507996082306, -0.042461175471544266, -0.015450406819581985, -0.03500641882419586, -0.06214277446269989, -0.0383245050907135, 0.046435944736003876, -0.07560601085424423, 0.013364278711378574, 0.023342855274677277, 0.05405820533633232, -0.025881100445985794, 0.06068144738674164, -0.08357544988393784, 0.09493788331747055, -0.1540430635213852, -0.03271956741809845, -0.025445878505706787, -0.041183918714523315, 0.1752462536096573, 0.06099751964211464, -0.015994304791092873, 0.15260063111782074, -0.17141541838645935, -0.058121129870414734, 0.15596486628055573, 0.008629098534584045, -0.09967197477817535, -0.003560945624485612, -0.09397093951702118, 0.1428760588169098, 0.08571921288967133, 0.2478504776954651, 0.12005335837602615, -0.22748184204101562, 0.055358242243528366, 0.12515293061733246, -0.14365963637828827, 0.10365243256092072, 0.07344598323106766, 0.005470725707709789, -0.18886831402778625, -0.06843198090791702, -0.06121627986431122, 0.1053021252155304, -0.08522345870733261, -0.0776243582367897, 0.09323626756668091, -0.05086790770292282, 0.24641476571559906, -0.028281206265091896, 0.06174173951148987, -0.026681531220674515, -0.1389324963092804, -0.01723906397819519, 0.060955192893743515, 0.05258452147245407, -0.024835573509335518, -0.25895482301712036, 0.13646544516086578, 0.048650871962308884, 0.025074828416109085, 0.004106190986931324, -0.05691491439938545, 0.016934165731072426, 0.1511998474597931, 0.020012924447655678, 0.13717477023601532, 0.027723990380764008, 0.0706823319196701, -0.006239562761038542, -0.10560829937458038, -0.04169593006372452, 0.061916545033454895, -0.08518962562084198, -0.06641357392072678, 0.011197872459888458, -0.06935211271047592, -0.11783787608146667, -0.12166737765073776, -0.026334572583436966, -0.02980303019285202, -0.07444227486848831, 0.02368103712797165, 0.06536602973937988, -0.06702698022127151, -0.0023908785078674555, 0.007125476840883493, -0.011537045240402222, 0.16434046626091003, 0.011393417604267597, -0.007796820718795061, 0.1328643560409546, -0.11533161997795105, 0.12461213022470474, 0.049438029527664185, -0.024806302040815353, -0.04662557691335678, 0.0014137453399598598, -0.057529181241989136, 0.029044216498732567, -0.04390640929341316, 0.02774495631456375, 0.20111067593097687, 0.02772962674498558, 0.11389166116714478, -0.0656520202755928, 0.04385066404938698, -0.007961965166032314, -0.009693224914371967, 0.018563594669103622, 0.07608018070459366, 0.07813210040330887, -0.1324140727519989, 0.02262016013264656, 0.22455167770385742, 0.1385764330625534, 0.18313980102539062, -0.010877152904868126, 0.06325667351484299, -0.04875868931412697, 0.027505528181791306, 0.024100203067064285, 0.10314226150512695, -0.10732068121433258, -0.0322517491877079, -0.025407759472727776, 0.023599207401275635, -0.08197105675935745, -0.1055799350142479, -0.090115025639534, 0.01222382951527834, -0.03125503659248352, -0.15570329129695892, 0.13300658762454987, -0.10451057553291321, 0.01802753657102585, 0.04692702740430832, -0.22163605690002441, 0.11530312895774841, 0.014291439205408096, -0.10303618758916855, 0.11281087249517441, -0.12051989883184433, -0.08699832111597061, -0.05777236074209213, -0.18658851087093353, 0.05280197039246559, 0.04673841595649719, 0.05166793242096901, -0.18521739542484283, 0.024835903197526932, 0.05545609071850777, 0.13426995277404785, -0.09743253141641617, -0.07142634689807892, -0.15038461983203888, 0.016068490222096443, -0.033661190420389175, -0.16029728949069977, -0.005609163548797369, -0.032781440764665604, -0.18849676847457886, -0.04539939761161804, -0.15086813271045685, -0.034627582877874374, 0.20464378595352173, 0.026907702907919884, 0.09480511397123337, -0.07926445454359055, 0.3802889585494995, -0.042039383202791214, -0.06146497279405594, -0.01321389526128769, -0.07072482258081436, 0.02512686513364315, 0.13271741569042206, 0.0036099457647651434, -0.017886579036712646, -0.0037857077550143003, 0.0024592927657067776, -0.06234965845942497, -0.13400450348854065, 0.0028710351325571537, 0.03905198723077774, 0.1874423623085022, 0.004639793653041124, 0.06659388542175293, 0.03133883699774742, 0.057546284049749374, 0.07748064398765564, 0.030926106497645378, 0.0011591583024710417, -0.01591806672513485, 0.06604493409395218, -0.11684755235910416, 0.042466625571250916, -0.030429253354668617, -0.10143838077783585, -0.013183288276195526, 0.07950251549482346, 0.12755028903484344, 0.17849206924438477, -0.04790908098220825, 0.17489230632781982, 0.13580141961574554, 0.16576050221920013, 0.049315933138132095, -0.020801831036806107, -0.08773037046194077, -0.06118565797805786, 0.004774159751832485, -0.031952597200870514, 0.04869702458381653, 0.3231290578842163, 0.037619613111019135, -0.09036035090684891, 0.11149907857179642, 0.009480619803071022, 0.05359881371259689, 0.022797370329499245, -0.11162138730287552, 0.11170321702957153, 0.07968773692846298, -0.06341761350631714, -0.07602835446596146, 0.16758501529693604, -0.1109386757016182, -0.26646625995635986, -0.11410990357398987, -0.012305386364459991, 0.07903840392827988, 0.005651174578815699, 0.05498376116156578, -0.11829282343387604, -0.16034497320652008, -0.034191906452178955, 0.1335442066192627, -0.3077351450920105, 0.2065143585205078, -0.0198091771453619, 0.06707923114299774, -0.039657969027757645, -0.07026876509189606, 0.09694647043943405, 0.13174086809158325, 0.29124146699905396, 0.01396956667304039, 0.04841272905468941, -0.15176129341125488, -0.0976925864815712, 0.0018439020495861769, 0.015482662245631218, -0.02563396655023098, 0.028520405292510986, -0.0540912002325058, 0.008404579944908619, -0.018086453899741173, 0.2102297693490982, -0.11316607892513275, 0.004344627261161804, -0.06968966871500015, -0.11707738786935806, 0.19409789144992828, -0.07178345322608948, -0.04543264955282211, -0.14959357678890228, -0.15512511134147644, -0.004174166824668646, -0.02413962036371231, -0.019664527848362923, -0.17603960633277893, -0.18804074823856354, -0.05204557999968529, -0.005645004566758871, -0.003464865731075406, 0.05867868289351463, -0.07517234236001968, -0.04805335775017738, 0.1009904220700264, -0.07743175327777863, -0.056063808500766754, -0.1103200614452362, 0.1391381323337555, 0.06248528137803078, 0.16743235290050507, 0.05907081440091133, 0.0006117874872870743, 0.11471151560544968, -0.02913086675107479, 0.11103474348783493, -0.11291708797216415, -0.17145049571990967, -0.08334989100694656, -0.018775060772895813, 0.09519003331661224, -0.04789286106824875, 0.0028788831550627947, 0.2550160884857178, 0.14880181849002838, -0.0897710770368576, 0.27680760622024536, 0.04414956644177437, -0.09375058114528656, -0.18432219326496124, -0.15961645543575287, 0.03759992495179176, 0.060025621205568314, 0.13095876574516296, -0.057205069810152054, -0.08483537286520004, -0.08492398262023926, -0.07478608191013336, -0.13140805065631866, -0.24232175946235657, -0.030598774552345276, 0.22874866425991058, 0.08656918257474899, 0.08219650387763977, -0.012482990510761738, -0.01186054851859808, 0.00526038184762001, 0.02680150233209133, 0.12018456310033798, -0.13341329991817474, 0.11107480525970459, 0.022198403254151344, 0.044267985969781876, 0.009712530300021172, 0.07929777354001999, 0.03375575691461563, -0.003218587953597307, -0.0006439819699153304, -0.0988350659608841, -0.2596651017665863, 0.0816885456442833, -0.01623627357184887, -0.09960969537496567, 0.014988959766924381, 0.02061903104186058, -0.2089255303144455, 0.011128270998597145, -0.019883770495653152, -0.03150356933474541, -0.06483490765094757, -0.10664787143468857, -0.056551624089479446, 0.04928823933005333, 0.10853826254606247, 0.011660109274089336, 0.05354316532611847, -0.0404130220413208, 0.07917837053537369, 0.0826287642121315, 0.15132710337638855, 0.06795957684516907, -0.190711110830307, -0.10953907668590546, -0.0414445661008358, 0.12121522426605225, -0.12505418062210083, 0.036917757242918015, 0.053161121904850006, -0.016534561291337013, 0.14621229469776154, 0.1070784479379654, -0.07452095299959183, 0.11915595084428787, 0.08904775977134705, -0.04094788804650307, -0.23367151618003845, -0.07120766490697861, 0.11133213341236115, 0.07195597887039185, -0.03961895406246185, 0.018120890483260155, -0.04960581287741661, -0.013980977237224579, 0.048759616911411285, -0.0538676381111145, -0.07230538129806519, 0.004421027842909098, 0.1247575581073761, 0.1029362753033638, -0.04655474051833153, 0.01296416949480772, 0.037371400743722916, 0.003788623260334134, 0.04730486497282982, 0.0407949760556221, -0.08269952982664108, -0.04124005511403084, 0.02782733179628849, 0.37552911043167114, -0.010165480896830559, -0.020456433296203613, 0.018555615097284317, -0.19949445128440857, 0.09135842323303223, 0.13205479085445404, 0.04697350412607193, 0.004247748292982578, -0.08139242231845856, 0.026877427473664284, -0.010625290684401989, 0.09936143457889557, -0.07806670665740967, -0.05493134260177612, -0.21631066501140594, -0.025010565295815468, 0.017490221187472343, 0.24077683687210083, -0.08458559215068817, -0.12801732122898102, -0.20628872513771057, 0.13128381967544556, -0.11333390325307846, -0.03695881739258766, -0.024473199620842934, 0.03926658630371094, -0.01989821158349514, 0.06291737407445908, -0.0710630789399147, 0.006373001262545586, -0.11024709790945053, 0.055267609655857086, 0.04204455390572548, 0.1229788213968277, 0.014207782223820686, 0.02016810141503811, 0.05822525918483734, -0.01837925612926483, 0.07173580676317215, -0.06203491613268852, -0.04550490900874138, 0.14224006235599518, -0.020255116745829582, -0.04152837023139, -0.0483345128595829, -0.036874305456876755, 0.11981741338968277, -0.05059147998690605, -0.007141099311411381, -0.054929375648498535, -0.06906463205814362, 0.03462086617946625, -0.009175732731819153, -0.008798843249678612, 0.06801853328943253, 0.04024988040328026, -0.026994358748197556, 0.005263668950647116, 0.03447828069329262, -0.10330043733119965, -0.04955084249377251, 0.16955432295799255, -0.0749620869755745, 0.10274054110050201, -0.031069839373230934, 0.018015999346971512, 0.005847334861755371, -0.022399673238396645, -0.015360680408775806, -0.1457086056470871, -0.06137600541114807, -0.09489979594945908, 0.11565322428941727, 0.08146517723798752, 0.03358805552124977, 0.04274565726518631, 0.019532648846507072, -0.04414922371506691, -0.038583990186452866, 0.12961317598819733, 0.08133101463317871, 0.012996876612305641, 0.01137041300535202, 0.01941833831369877, -0.020302120596170425, 0.0028480992186814547, -0.01250747125595808, -0.07239153981208801, -0.05874783173203468, 0.09400010108947754, 0.1600283533334732, -0.06127211079001427, -0.13325586915016174, -0.020593497902154922, 0.04988488554954529, 0.0014717020094394684, -0.08777432143688202, 0.04833676666021347, 0.15805292129516602, -0.05623878911137581, 0.03216489031910896, -0.09984751045703888, -0.07263360917568207, -0.16060975193977356, -0.10029061883687973, -0.06092562898993492, -0.28350353240966797, 0.09752398729324341, 0.006392303854227066, -0.014731393195688725, 0.059529416263103485, 0.051305368542671204, -0.052508849650621414, 0.07068239152431488, -0.18146829307079315, -0.007054794579744339, 0.03497592359781265, -0.13212306797504425, 0.02475893869996071, -0.2378365397453308, 0.10198072344064713, -0.04623803123831749, -0.1519704908132553, -0.04004510119557381, 0.0641569048166275, -0.09540136158466339, -0.01822364516556263, -0.0475153923034668, -0.01922670193016529, 0.01624443754553795, -0.009348669089376926, -0.031147832050919533, 0.13716529309749603, 0.02827494591474533, -0.03268734738230705, 0.005254602525383234, 0.0223685409873724, 0.03955082967877388, -0.0969657450914383, -0.05986930429935455, 0.08311155438423157, -0.031056145206093788, 0.14728976786136627, 0.000341245875461027, 0.04181376099586487, -0.06758682429790497, 0.2593761384487152, 0.2023983597755432, -0.12479214370250702, 0.008118697442114353, -0.021801479160785675, 0.012670028023421764, -0.041751839220523834, 0.13110700249671936, 0.013386172242462635, 0.12186761200428009, -0.17513342201709747, -0.01036517322063446, -0.0818324014544487, -0.04501292482018471, 0.06702108681201935, 0.14714950323104858, 0.15742522478103638, 0.03436789661645889, -0.07328428328037262, 0.06722653657197952, -0.30119743943214417, 0.20540550351142883, -0.1346001923084259, -0.01498429011553526, -0.040251150727272034, -0.058389630168676376, 0.061147745698690414, 0.11309876292943954, 0.10832664370536804, -0.021150551736354828, -0.0905047357082367, -0.04486766457557678, -0.039378076791763306, -0.13019338250160217, -0.02718670479953289, 0.1654091775417328, 0.06799814850091934, 0.31520840525627136, -0.017577875405550003, 0.07702425122261047, 0.034410297870635986, 0.06451138854026794, 0.004519328009337187, 0.09537279605865479, 0.07960964739322662, -0.06345855444669724, -0.07373003661632538, -0.001637450186535716, 0.05033271387219429, 0.14567798376083374, -0.03826142102479935, -0.18691548705101013, 0.15858715772628784, 0.07192251086235046, -0.13762691617012024, -0.05777517706155777, 0.08409425616264343, -0.0739973932504654, 0.0550808347761631, 0.08115427941083908, 0.015876613557338715, -0.017793258652091026, -0.004664506763219833, 0.06074233725667, 0.024694660678505898, -0.02343848906457424, 0.003570882137864828, -0.08337053656578064, -0.04151543974876404, 0.07267895340919495, -0.0844460055232048, -0.20546193420886993, -0.0957019031047821, -0.07551700621843338, 0.030557552352547646, -0.0649830624461174, 0.12575586140155792, 0.1717868149280548, 0.0593598335981369, -0.03307248651981354, -0.10721943527460098, -0.035562749952077866, 0.07602505385875702, -0.044773899018764496, -0.09409699589014053 ]
null
null
diffusers
# frnkzane <Gallery /> ## Trigger words You should use `frnkzane` to trigger the image generation. ## Download model [Download](/Pfs2021Funny/Frank-Zane/tree/main) them in the Files & versions tab.
{"tags": ["text-to-image", "stable-diffusion", "lora", "diffusers", "template:sd-lora"], "widget": [{"text": "-", "output": {"url": "images/82642704991.jpg"}}], "base_model": "stabilityai/stable-diffusion-xl-base-1.0", "instance_prompt": "frnkzane"}
text-to-image
Pfs2021Funny/Frank-Zane
[ "diffusers", "text-to-image", "stable-diffusion", "lora", "template:sd-lora", "base_model:stabilityai/stable-diffusion-xl-base-1.0", "region:us" ]
2024-02-10T12:00:41+00:00
[]
[]
TAGS #diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #region-us
# frnkzane <Gallery /> ## Trigger words You should use 'frnkzane' to trigger the image generation. ## Download model Download them in the Files & versions tab.
[ "# frnkzane\n\n<Gallery />", "## Trigger words\n\nYou should use 'frnkzane' to trigger the image generation.", "## Download model\n\n\nDownload them in the Files & versions tab." ]
[ "TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #region-us \n", "# frnkzane\n\n<Gallery />", "## Trigger words\n\nYou should use 'frnkzane' to trigger the image generation.", "## Download model\n\n\nDownload them in the Files & versions tab." ]
[ 56, 9, 18, 14 ]
[ "passage: TAGS\n#diffusers #text-to-image #stable-diffusion #lora #template-sd-lora #base_model-stabilityai/stable-diffusion-xl-base-1.0 #region-us \n# frnkzane\n\n<Gallery />## Trigger words\n\nYou should use 'frnkzane' to trigger the image generation.## Download model\n\n\nDownload them in the Files & versions tab." ]
[ -0.08398307114839554, -0.09166717529296875, -0.005762984976172447, 0.05153873935341835, 0.07648049294948578, 0.0388760007917881, 0.08621177077293396, 0.045397866517305374, 0.152518630027771, 0.10905582457780838, 0.10933898389339447, 0.08414892852306366, 0.019768789410591125, 0.1860697716474533, -0.08993426710367203, -0.22171775996685028, 0.04442205652594566, 0.017714569345116615, -0.08399394154548645, 0.015255135484039783, 0.04169351980090141, -0.044029973447322845, 0.13654743134975433, -0.02213948965072632, -0.1424517184495926, 0.0300593264400959, 0.01824805699288845, -0.01651538535952568, 0.02257682941854, 0.10877587646245956, -0.027056386694312096, 0.13548128306865692, 0.05586153641343117, -0.1750134974718094, 0.07767274230718613, 0.023427508771419525, -0.07733748853206635, 0.03744605928659439, 0.025471162050962448, -0.03687523305416107, 0.2614181935787201, -0.07617707550525665, -0.11604175716638565, -0.01993773691356182, -0.0551360659301281, -0.08183906227350235, -0.010965543799102306, 0.05727313831448555, 0.03413822874426842, -0.03784015774726868, 0.030257390812039375, -0.023301495239138603, 0.015354856848716736, 0.0847763866186142, 0.21961821615695953, -0.17699600756168365, -0.08078017830848694, 0.22102898359298706, 0.051621031016111374, 0.1360037475824356, -0.08934774994850159, 0.1658376157283783, 0.08033180981874466, -0.06468796730041504, -0.014798265881836414, -0.06342735141515732, 0.12860922515392303, -0.07886306196451187, -0.07432930916547775, 0.0287062618881464, 0.2209763079881668, 0.075395368039608, 0.0018145984504371881, -0.08402889221906662, -0.08248303830623627, 0.16766563057899475, -0.1167488768696785, -0.03981129080057144, -0.010600115172564983, 0.002972609829157591, 0.07760703563690186, -0.18752487003803253, -0.07586734741926193, -0.056786783039569855, -0.024815570563077927, 0.2646929621696472, 0.014140153303742409, 0.08312833309173584, 0.02693871781229973, 0.05136329308152199, -0.14731992781162262, -0.1252782791852951, -0.005962550174444914, -0.098052017390728, 0.09297104179859161, 0.09265494346618652, -0.019085172563791275, -0.049377840012311935, 0.16057172417640686, 0.0920126810669899, -0.0739879310131073, -0.0007815933786332607, 0.004240589682012796, 0.09822983294725418, 0.0033248737454414368, -0.13224665820598602, -0.04855125769972801, -0.06913188844919205, 0.06162456423044205, 0.009457331150770187, 0.01345011219382286, -0.08494007587432861, -0.12444338947534561, -0.029445895925164223, -0.13166120648384094, -0.0359884612262249, 0.006645821966230869, 0.025833051651716232, 0.034946292638778687, -0.047291625291109085, 0.14246651530265808, 0.006605442147701979, -0.07047460973262787, 0.024578552693128586, -0.042782824486494064, 0.1989823877811432, 0.0612509585916996, 0.006554055493324995, 0.07694245129823685, 0.03404167667031288, -0.050151802599430084, -0.06445079296827316, -0.025004122406244278, -0.006267961580306292, -0.003958242014050484, -0.14927250146865845, 0.059212375432252884, -0.13080435991287231, -0.15113528072834015, 0.01756102591753006, 0.0857514888048172, -0.0768970176577568, 0.007995689287781715, -0.012034091167151928, -0.032720908522605896, 0.04316890984773636, 0.022269176319241524, 0.05632936581969261, -0.04508678987622261, 0.08335316926240921, 0.0007445168448612094, 0.16762056946754456, -0.17693406343460083, -0.010716415010392666, 0.003319954965263605, 0.05175623670220375, -0.1199849471449852, 0.04455070570111275, -0.09820147603750229, -0.01976858451962471, -0.03729226812720299, -0.009197329171001911, -0.13720841705799103, 0.03961219638586044, -0.004731354303658009, 0.1978560835123062, -0.022371018305420876, -0.059183575212955475, 0.0818350538611412, -0.1491592973470688, -0.03544284775853157, 0.03504721820354462, 0.015554976649582386, 0.08342365175485611, 0.02557164803147316, 0.21825462579727173, 0.061357103288173676, -0.26545628905296326, 0.09911289066076279, 0.13515138626098633, -0.1342715322971344, -0.02018716000020504, 0.030096787959337234, 0.1286211907863617, 0.03286449983716011, 0.08059936761856079, -0.10457471013069153, 0.13669244945049286, -0.10599872469902039, 0.010859078727662563, -0.02219518832862377, -0.0742437094449997, -0.03489824756979942, 0.03223898261785507, 0.04825403541326523, 0.032222457230091095, 0.000572830147575587, -0.0451098307967186, 0.09438039362430573, -0.03636449575424194, -0.0026552672497928143, -0.02029496245086193, 0.21101424098014832, -0.0741635337471962, 0.03328485041856766, -0.05963290482759476, -0.08024284243583679, 0.017734616994857788, 0.11002899706363678, 0.01804329641163349, 0.06089482828974724, 0.096822589635849, 0.022627996280789375, -0.05576995760202408, 0.032747864723205566, 0.07154390960931778, -0.04641938582062721, 0.005002670455724001, -0.18897207081317902, 0.019764773547649384, -0.058559663593769073, 0.11081118881702423, -0.3000899851322174, 0.01281528640538454, 0.08581238240003586, 0.10081357508897781, 0.021644117310643196, 0.004279403481632471, -0.028932293877005577, -0.09700433164834976, -0.06970980018377304, 0.012994767166674137, 0.07914391160011292, 0.00416305847465992, 0.020529314875602722, 0.13998568058013916, 0.010520211420953274, 0.10277225077152252, 0.12397260218858719, -0.05375567078590393, 0.0067186807282269, -0.1646527200937271, -0.003695930354297161, 0.031581390649080276, 0.07278769463300705, -0.0040429458022117615, -0.17437204718589783, -0.023840665817260742, 0.06768222153186798, -0.05108863115310669, 0.08165310323238373, 0.12251481413841248, -0.041971880942583084, -0.07543792575597763, 0.08268719166517258, 0.1794194132089615, -0.045386962592601776, 0.07625620812177658, 0.09571827203035355, -0.019278593361377716, 0.17378292977809906, -0.009652574546635151, -0.11266455799341202, 0.004228921141475439, 0.09058746695518494, -0.008161023259162903, 0.19151629507541656, 0.030815012753009796, -0.03152984380722046, 0.0452735498547554, -0.11494425684213638, 0.022035418078303337, -0.0848328024148941, -0.06984551250934601, 0.005156128201633692, -0.056242916733026505, 0.1573704034090042, 0.1416260153055191, -0.09888119995594025, 0.07508405297994614, -0.07768787443637848, -0.09868694096803665, -0.04202356934547424, -0.018916361033916473, 0.01710052788257599, 0.05831465497612953, 0.032177962362766266, -0.16771380603313446, -0.10477888584136963, -0.02879577875137329, -0.13997213542461395, -0.017083602026104927, -0.0038792237173765898, -0.049913786351680756, -0.007657474838197231, -0.032457999885082245, 0.00695737823843956, 0.01782592199742794, -0.03124489076435566, -0.04270700737833977, -0.0342814140021801, -0.10789292305707932, -0.05593447387218475, -0.032537396997213364, -0.12991920113563538, 0.03405705839395523, 0.11074364185333252, -0.1021340936422348, 0.17599746584892273, 0.041263241320848465, 0.06393719464540482, 0.08008524775505066, 0.02608705684542656, 0.18106183409690857, -0.12608805298805237, 0.12088057398796082, 0.20086389780044556, 0.04412181302905083, 0.04744414612650871, 0.057053398340940475, 0.025079019367694855, -0.10505852848291397, 0.013544620014727116, 0.031554851680994034, -0.11562035232782364, -0.0320277065038681, -0.0850689634680748, -0.11830321699380875, -0.024997098371386528, 0.02159884199500084, 0.032596223056316376, 0.15038269758224487, 0.1317838579416275, 0.0011921600671485066, -0.09470164030790329, 0.04934589937329292, 0.02831110917031765, 0.14204740524291992, -0.01992212049663067, 0.05540420860052109, -0.017398744821548462, -0.022199686616659164, 0.14703822135925293, -0.01250643189996481, 0.2099105715751648, -0.0014124746667221189, 0.0614311583340168, 0.02690058946609497, 0.14007721841335297, 0.0840621143579483, 0.07891150563955307, -0.06026202440261841, -0.037654392421245575, -0.03418150544166565, -0.09867727756500244, 0.08033217489719391, 0.07325012236833572, -0.005524453707039356, -0.08941742777824402, 0.02981187030673027, 0.028583243489265442, -0.013777174986898899, 0.07389141619205475, -0.0054756393656134605, -0.20603156089782715, 0.023449642583727837, -0.007570798508822918, 0.12784531712532043, -0.05974565073847771, 0.020889587700366974, 0.17011518776416779, -0.041796132922172546, 0.025200974196195602, -0.04881971329450607, 0.06340456753969193, 0.028675822541117668, -0.0648888647556305, -0.03403591364622116, 0.12905281782150269, -0.010140589438378811, -0.00787516962736845, -0.047550275921821594, 0.1161934956908226, -0.006650297902524471, -0.043963752686977386, -0.0035349270328879356, -0.03650956600904465, 0.06049526110291481, 0.14274121820926666, 0.2096499800682068, 0.018468940630555153, 0.057226717472076416, -0.0846015065908432, -0.08666904270648956, 0.019968636333942413, 0.15316890180110931, -0.0629149004817009, -0.005691306199878454, 0.0248117633163929, -0.02354806661605835, 0.015702249482274055, 0.06346290558576584, -0.18787391483783722, -0.12337907403707504, 0.018866389989852905, 0.06420493870973587, -0.07361220568418503, -0.02752663567662239, -0.04749513417482376, -0.09625916182994843, 0.026944929733872414, 0.12135746330022812, -0.04764917865395546, -0.1284584254026413, 0.0021982265170663595, 0.043172791600227356, -0.04441724345088005, 0.07373392581939697, -0.08843467384576797, 0.051234833896160126, -0.0957656055688858, -0.15289559960365295, 0.04254841059446335, -0.07048595696687698, -0.035354964435100555, -0.020784834399819374, 0.12200184166431427, 0.040458742529153824, -0.05860956385731697, 0.04386099427938461, 0.008579191751778126, 0.04682830348610878, -0.10012032091617584, 0.04583793133497238, -0.07358374446630478, -0.10841204971075058, -0.02186979167163372, -0.056517090648412704, 0.026811042800545692, -0.048192691057920456, 0.005340317729860544, 0.05335390567779541, 0.2545875906944275, -0.06842255592346191, 0.029949838295578957, 0.0898885726928711, -0.004859419539570808, -0.22800742089748383, -0.02866097167134285, -0.0013419364113360643, -0.04861225187778473, -0.01633519120514393, -0.14655669033527374, 0.09586051106452942, 0.08630379289388657, -0.029396535828709602, 0.3649831712245941, -0.29623332619667053, -0.07046084105968475, 0.02328895777463913, 0.14475461840629578, 0.289108544588089, -0.21819116175174713, -0.04935179278254509, -0.004798483569175005, -0.14669857919216156, 0.14790472388267517, -0.05034889653325081, 0.0368422232568264, -0.01920609176158905, -0.09108711034059525, 0.011030382476747036, 0.0021097164135426283, 0.15759147703647614, -0.036965012550354004, 0.07399316877126694, -0.07842084020376205, -0.0327724851667881, 0.08266747742891312, 0.0060502393171191216, -0.0026551177725195885, -0.1380167156457901, -0.004420367069542408, -0.14572839438915253, -0.04536201059818268, -0.04861371964216232, 0.11863407492637634, 0.0065360721200704575, -0.04617959260940552, -0.06858285516500473, 0.030159572139382362, 0.011199764907360077, 0.068747378885746, 0.08679579943418503, -0.07219696789979935, -0.03614078089594841, 0.2740321457386017, -0.14453008770942688, 0.020937737077474594, -0.03953130915760994, -0.07611948996782303, -0.023050954565405846, 0.09129999577999115, -0.15887393057346344, -0.0593784898519516, 0.11071521788835526, 0.07765737175941467, 0.06342332810163498, 0.006685157306492329, -0.011547381058335304, 0.05736677721142769, 0.13299696147441864, -0.13256819546222687, -0.11673139035701752, -0.07538159191608429, -0.20686937868595123, 0.17210902273654938, -0.0037242670077830553, 0.12407397478818893, -0.061752717941999435, 0.05449514091014862, -0.012950411066412926, -0.01588442549109459, -0.05419392138719559, 0.027201920747756958, 0.10287017375230789, -0.044059284031391144, -0.09045757353305817, 0.03487062454223633, -0.03431049734354019, -0.044529374688863754, -0.12510298192501068, 0.05495172739028931, -0.08377455174922943, -0.0566694401204586, -0.00899395439773798, 0.013547784648835659, -0.16996552050113678, -0.00668598897755146, -0.004357875324785709, -0.0825265645980835, -0.006285546347498894, 0.1349315047264099, 0.0760745108127594, -0.1117541491985321, 0.05011749267578125, 0.07872752100229263, -0.03363582864403725, 0.04716421663761139, 0.07707399874925613, 0.0667911171913147, -0.12076831609010696, -0.17715631425380707, 0.006243426818400621, 0.03794350102543831, -0.12357693910598755, -0.07328596711158752, -0.07570179551839828, -0.015523141250014305, -0.03335389867424965, 0.06005435436964035, -0.08647569268941879, -0.037917494773864746, -0.04768732562661171, -0.09058095514774323, -0.05857040733098984, 0.014717871323227882, -0.06646434217691422, 0.0177783016115427, 0.028300484642386436, 0.02235754020512104, -0.06726571917533875, -0.07958477735519409, 0.02510879375040531, -0.08587805926799774, 0.06062481179833412, 0.11504735052585602, -0.03929037228226662, 0.04325616732239723, -0.24670125544071198, -0.0021833423525094986, 0.08537547290325165, 0.07517755776643753, -0.04657271131873131, 0.11024845391511917, 0.08690627664327621, -0.026647217571735382, 0.024828052148222923, -0.040618281811475754, -0.06236005201935768, -0.10016173869371414, 0.0731220617890358, -0.10098651796579361, -0.03733863681554794, 0.0009385923040099442, 0.025231562554836273, 0.11051236093044281, 0.09950345009565353, 0.05807044357061386, -0.05588962510228157, 0.017165835946798325, -0.0746379867196083, 0.02820347435772419, 0.06490529328584671, -0.09926339238882065, 0.03391892835497856, -0.10283079743385315, 0.002984498394653201, 0.008259404450654984, 0.14828981459140778, 0.0005595539114437997, -0.12561574578285217, -0.032341573387384415, -0.029844919219613075, 0.04269535094499588, 0.008439398370683193, 0.2445623278617859, 0.15351895987987518, 0.10780509561300278, -0.08663161844015121, 0.10161928832530975, 0.12821932137012482, -0.04356294125318527, 0.0032333512790501118, 0.15005925297737122, -0.030573498457670212, 0.0796470120549202, 0.09381639957427979, 0.0074410876259207726, 0.12513750791549683, 0.12349089980125427, -0.09912265837192535, 0.045490968972444534, -0.03321513533592224, -0.007861756719648838, 0.19908569753170013, -0.07817161083221436, -0.000047227204049704596, 0.08600617945194244, -0.007766894996166229, -0.08387412130832672, -0.21986089646816254, -0.07248208671808243, -0.20195724070072174, 0.05382970720529556, -0.05336211994290352, 0.023255566135048866, 0.056554146111011505, 0.07369216531515121, 0.019843699410557747, 0.07889843732118607, -0.018434390425682068, -0.09652926027774811, 0.1678299456834793, -0.03448814898729324, -0.10074157267808914, -0.06557357311248779, -0.04725533723831177, 0.03241676464676857, -0.04904784634709358, -0.0449664369225502, 0.06179118528962135, 0.044623129069805145, 0.05231096222996712, -0.0218571275472641, -0.05503050237894058, -0.033469367772340775, 0.03321541100740433, -0.020731588825583458, 0.09876465797424316, 0.017991311848163605, 0.008609618060290813, -0.025622360408306122, 0.10525210201740265, 0.019849678501486778, -0.013652944006025791, -0.021181046962738037, -0.03651556745171547, -0.07404398173093796, 0.1340884119272232, -0.08226649463176727, -0.03736012801527977, -0.042704880237579346, 0.2876843512058258, 0.18367619812488556, -0.14911653101444244, 0.030701318755745888, 0.05409007892012596, 0.019024765118956566, 0.03404918313026428, 0.05195316672325134, 0.025807401165366173, 0.1807972490787506, -0.018998373299837112, -0.03895829617977142, -0.10926982760429382, -0.0221951212733984, -0.04206828400492668, -0.06714801490306854, 0.07916603982448578, -0.06873102486133575, -0.08740074932575226, 0.10410725325345993, -0.15210257470607758, -0.11193110793828964, 0.14011195302009583, -0.04990033432841301, -0.0003625802928581834, -0.10852516442537308, -0.10367453098297119, 0.08246022462844849, -0.017078382894396782, -0.14215780794620514, -0.043967075645923615, -0.057063691318035126, 0.009611520916223526, -0.12138444185256958, -0.07219792902469635, 0.019047725945711136, -0.1498221457004547, 0.10587003082036972, -0.06742474436759949, 0.04308045655488968, -0.01715162768959999, -0.049471817910671234, -0.00442233681678772, -0.01096862182021141, 0.026565106585621834, -0.13456274569034576, -0.09274687618017197, 0.06855456531047821, -0.0021151150576770306, 0.036333177238702774, 0.01475280337035656, -0.1063639298081398, 0.04245484992861748, 0.17529039084911346, -0.06575138866901398, -0.03499792143702507, 0.053860854357481, -0.1545848399400711, 0.047371406108140945, 0.04599948227405548, 0.05035135895013809, -0.04104877635836601, -0.03901709243655205, 0.07319687306880951, 0.11958285421133041, 0.007882798090577126, 0.021352214738726616, -0.048452895134687424, -0.04778999462723732, -0.06135307252407074, 0.0027341400273144245, -0.07397040724754333, -0.013705396093428135, -0.17130158841609955, 0.055647194385528564, 0.013558943755924702, 0.06359551846981049, 0.2016405612230301, -0.0021455897949635983, 0.0018155508441850543, -0.24731606245040894, 0.043467387557029724, 0.07766427844762802, -0.07393824309110641, 0.011831531301140785 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Whisper Small Tr Cv11 - tgrhn This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the Common Voice 11.0 dataset. It achieves the following results on the evaluation set: - Loss: 0.4045 - Wer: 104.3596 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 64 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - training_steps: 5000 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:--------:| | 0.0776 | 2.04 | 1000 | 0.2936 | 70.0976 | | 0.0114 | 4.07 | 2000 | 0.3641 | 83.6851 | | 0.0038 | 7.03 | 3000 | 0.3743 | 94.4617 | | 0.0013 | 9.07 | 4000 | 0.4023 | 114.4720 | | 0.0011 | 12.02 | 5000 | 0.4045 | 104.3596 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.1.1 - Datasets 2.16.1 - Tokenizers 0.15.1
{"language": ["tr"], "license": "apache-2.0", "tags": ["whisper-event", "generated_from_trainer"], "datasets": ["mozilla-foundation/common_voice_11_0"], "metrics": ["wer"], "base_model": "openai/whisper-small", "model-index": [{"name": "Whisper Small Tr Cv11 - tgrhn", "results": [{"task": {"type": "automatic-speech-recognition", "name": "Automatic Speech Recognition"}, "dataset": {"name": "Common Voice 11.0", "type": "mozilla-foundation/common_voice_11_0", "config": "tr", "split": "test", "args": "tr"}, "metrics": [{"type": "wer", "value": 104.359626480413, "name": "Wer"}]}]}]}
automatic-speech-recognition
tgrhn/whisper-small-tr-cv11
[ "transformers", "tensorboard", "whisper", "automatic-speech-recognition", "whisper-event", "generated_from_trainer", "tr", "dataset:mozilla-foundation/common_voice_11_0", "base_model:openai/whisper-small", "license:apache-2.0", "model-index", "endpoints_compatible", "region:us" ]
2024-02-10T12:06:03+00:00
[]
[ "tr" ]
TAGS #transformers #tensorboard #whisper #automatic-speech-recognition #whisper-event #generated_from_trainer #tr #dataset-mozilla-foundation/common_voice_11_0 #base_model-openai/whisper-small #license-apache-2.0 #model-index #endpoints_compatible #region-us
Whisper Small Tr Cv11 - tgrhn ============================= This model is a fine-tuned version of openai/whisper-small on the Common Voice 11.0 dataset. It achieves the following results on the evaluation set: * Loss: 0.4045 * Wer: 104.3596 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 1e-05 * train\_batch\_size: 64 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 500 * training\_steps: 5000 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.37.2 * Pytorch 2.1.1 * Datasets 2.16.1 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 5000\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.1\n* Datasets 2.16.1\n* Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #tensorboard #whisper #automatic-speech-recognition #whisper-event #generated_from_trainer #tr #dataset-mozilla-foundation/common_voice_11_0 #base_model-openai/whisper-small #license-apache-2.0 #model-index #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 5000\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.1\n* Datasets 2.16.1\n* Tokenizers 0.15.1" ]
[ 95, 130, 4, 30 ]
[ "passage: TAGS\n#transformers #tensorboard #whisper #automatic-speech-recognition #whisper-event #generated_from_trainer #tr #dataset-mozilla-foundation/common_voice_11_0 #base_model-openai/whisper-small #license-apache-2.0 #model-index #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 5000\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.1\n* Datasets 2.16.1\n* Tokenizers 0.15.1" ]
[ -0.10236019641160965, 0.09202783554792404, -0.0033191274851560593, 0.046530142426490784, 0.10148610174655914, -0.012793525122106075, 0.1304219365119934, 0.14117340743541718, -0.06279175728559494, 0.07614465057849884, 0.08320732414722443, 0.07758215069770813, 0.07367368042469025, 0.12461961060762405, -0.01168301422148943, -0.318483829498291, 0.027777493000030518, -0.0005648660007864237, -0.08032963424921036, 0.1045568659901619, 0.10397099703550339, -0.09769909828901291, 0.01856500655412674, 0.025974562391638756, -0.09667030721902847, 0.012252365238964558, -0.014998088590800762, -0.06743492931127548, 0.10534759610891342, 0.03140590712428093, 0.0622343011200428, 0.022382987663149834, 0.08202429860830307, -0.26825079321861267, 0.018464043736457825, 0.04965431988239288, 0.061495091766119, 0.059738125652074814, 0.077583909034729, -0.007867119275033474, 0.08364354074001312, -0.03317059949040413, 0.07552121579647064, 0.06426810473203659, -0.10134027898311615, -0.3000025153160095, -0.06568681448698044, 0.03425640985369682, 0.11706816405057907, 0.0905766636133194, -0.03787922486662865, 0.07007187604904175, -0.0482703372836113, 0.09343113750219345, 0.21026284992694855, -0.22609275579452515, -0.08150330185890198, -0.062115564942359924, 0.057351354509592056, 0.04624754935503006, -0.11265923082828522, -0.012456650845706463, 0.03674397990107536, 0.03643554076552391, 0.09020017832517624, 0.008766194805502892, -0.0052859121933579445, -0.014670295640826225, -0.135783389210701, -0.05615510419011116, 0.14963427186012268, 0.07848480343818665, -0.04422379285097122, -0.09799770265817642, -0.008883307687938213, -0.14293546974658966, -0.048962727189064026, 0.009128818288445473, 0.02737552300095558, -0.0206894613802433, -0.09187348932027817, -0.008038969710469246, -0.0844089463353157, -0.09017603099346161, 0.037492021918296814, 0.16187770664691925, 0.03904031589627266, -0.03742365911602974, -0.014190562069416046, 0.08285592496395111, 0.052716903388500214, -0.14294631779193878, -0.015276308171451092, 0.043652746826410294, -0.08426227420568466, -0.020546268671751022, -0.032131120562553406, -0.04688543826341629, 0.023743800818920135, 0.13224221765995026, 0.015637874603271484, 0.0991254448890686, 0.0026328591629862785, 0.0368819497525692, -0.08588755130767822, 0.17111344635486603, -0.04652777686715126, -0.024898074567317963, -0.0375559963285923, 0.12028344720602036, -0.006334271281957626, -0.02198944054543972, -0.05714989826083183, 0.049571145325899124, 0.07985355705022812, 0.03918414190411568, -0.010806885547935963, 0.02974598854780197, -0.0680103451013565, -0.02346971072256565, -0.10245079547166824, -0.11767216771841049, 0.013579430989921093, 0.026810938492417336, -0.062025755643844604, -0.009136543609201908, 0.009519550018012524, 0.03906211629509926, -0.027133336290717125, 0.047813206911087036, -0.05356348678469658, -0.0009041558369062841, -0.07850776612758636, -0.0859900638461113, 0.026762565597891808, -0.02008073218166828, 0.007477226201444864, -0.07547423988580704, -0.10774131864309311, -0.05340544506907463, 0.07311578840017319, -0.039471909403800964, -0.08335515856742859, -0.09010834246873856, -0.07534012943506241, 0.044057730585336685, -0.01791391707956791, 0.13390938937664032, -0.0505090095102787, 0.09606050699949265, 0.020629001781344414, 0.04153947904706001, 0.04811835289001465, 0.0693422332406044, -0.04173947870731354, 0.049389030784368515, -0.12437039613723755, 0.10627431422472, -0.1055445745587349, 0.07885056734085083, -0.14345631003379822, -0.09819840639829636, -0.015423896722495556, 0.009167499840259552, 0.10454896092414856, 0.11281250417232513, -0.1747635006904602, -0.09704166650772095, 0.17520949244499207, -0.07750248908996582, -0.10468058288097382, 0.15086989104747772, -0.02381780557334423, 0.04726492241024971, 0.05048668757081032, 0.23591069877147675, 0.09212446957826614, -0.08549701422452927, 0.03149009868502617, -0.04693925753235817, 0.10418549180030823, 0.04423040896654129, 0.09674682468175888, -0.05544838681817055, -0.0004150010645389557, -0.0024821944534778595, -0.02063112147152424, 0.08336146920919418, -0.08183585107326508, -0.08571905642747879, -0.008026953786611557, -0.08007896691560745, 0.009719686582684517, 0.049083005636930466, 0.026326030492782593, -0.09899546205997467, -0.11214683949947357, 0.042408280074596405, 0.10836055874824524, -0.11365201324224472, 0.02878563292324543, -0.10952366143465042, 0.02677920088171959, -0.015318106859922409, -0.008905011229217052, -0.1306179165840149, 0.02381608635187149, 0.0350978784263134, -0.06573498994112015, 0.04456170275807381, -0.03676455840468407, 0.08720792829990387, 0.0274334903806448, -0.029255740344524384, -0.07446552067995071, -0.06777078658342361, 0.0036882951389998198, -0.07950099557638168, -0.220982626080513, -0.06838496774435043, -0.03563252091407776, 0.17288534343242645, -0.2098635584115982, 0.022873081266880035, 0.02621651440858841, 0.1079993024468422, 0.034169554710388184, -0.052471715956926346, 0.017092028632760048, 0.051189590245485306, -0.0052718594670295715, -0.06982274353504181, 0.031622808426618576, 0.02844284661114216, -0.1453048586845398, 0.046090010553598404, -0.15278910100460052, 0.08386672288179398, 0.08568266779184341, -0.02711968682706356, -0.05560824275016785, -0.048804666846990585, -0.06063627451658249, -0.06561755388975143, -0.014897813089191914, -0.024716423824429512, 0.2342863827943802, 0.013889772817492485, 0.11117532104253769, -0.08252591639757156, -0.04894949495792389, 0.020577989518642426, -0.02662750706076622, -0.01969243586063385, 0.1395094096660614, 0.012574717402458191, -0.06121562048792839, 0.08184077590703964, 0.05528946965932846, -0.07667964696884155, 0.1882329136133194, -0.07867791503667831, -0.10255169868469238, -0.02115098014473915, 0.030811134725809097, 0.03538952022790909, 0.10317818820476532, -0.18114560842514038, -0.010492757894098759, 0.016720173880457878, 0.014279140159487724, 0.03716937452554703, -0.19544069468975067, -0.00860498659312725, 0.05030783265829086, -0.07473152875900269, -0.045545365661382675, -0.013473838567733765, -0.012527568265795708, 0.07722657918930054, 0.000894813914783299, -0.05436088144779205, -0.004253517836332321, -0.042904797941446304, -0.09499171376228333, 0.17352114617824554, -0.08989284187555313, -0.14677250385284424, -0.1418597400188446, -0.023525413125753403, 0.015272499062120914, -0.005420489702373743, 0.055436741560697556, -0.1129867359995842, -0.02768348902463913, -0.08052103966474533, 0.04355698078870773, -0.055279411375522614, 0.02329115942120552, 0.04177306592464447, 0.0028607200365513563, 0.09844128042459488, -0.09912144392728806, 0.014445281587541103, -0.016620034351944923, -0.030506305396556854, -0.0032933317124843597, 0.004656411241739988, 0.08012975007295609, 0.1688782125711441, 0.04790296033024788, 0.02961244434118271, -0.03865516185760498, 0.16561168432235718, -0.11404041945934296, -0.02651161141693592, 0.1367310881614685, -0.02040199004113674, 0.04460179805755615, 0.13117314875125885, 0.043804652988910675, -0.0848354771733284, 0.033354368060827255, 0.03670158609747887, -0.024900805205106735, -0.2359033077955246, -0.014620772562921047, -0.06149493530392647, -0.00938192568719387, 0.09826061129570007, 0.030170772224664688, 0.00833301991224289, 0.030805565416812897, -0.016872912645339966, -0.025913381949067116, 0.036898765712976456, 0.06784474849700928, 0.06498929113149643, 0.01723150722682476, 0.09945761412382126, -0.015791654586791992, -0.04640856385231018, 0.008395764976739883, 0.005996244493871927, 0.22698338329792023, 0.018521009013056755, 0.1927412897348404, 0.05936584621667862, 0.1162291169166565, 0.007316288538277149, 0.05110901966691017, 0.0159190334379673, -0.0159160066395998, 0.01878068596124649, -0.052910879254341125, -0.04357744753360748, 0.045526910573244095, 0.09076124429702759, 0.055047858506441116, -0.11255090683698654, 0.012799283489584923, 0.016730789095163345, 0.3734784424304962, 0.05768047645688057, -0.27355965971946716, -0.10250972956418991, 0.01940801739692688, -0.10645891726016998, -0.04111572355031967, 0.03769851475954056, 0.14610138535499573, -0.0920010358095169, 0.03809354826807976, -0.06596805155277252, 0.09654034674167633, -0.06673365086317062, 0.014058257453143597, 0.055547747761011124, 0.09149804711341858, -0.006447901017963886, 0.05919172614812851, -0.2668428122997284, 0.2997833788394928, -0.020411711186170578, 0.10169021785259247, -0.03641311451792717, 0.03781330958008766, 0.039138954132795334, -0.062315672636032104, 0.07787660509347916, -0.0036442868877202272, -0.10876596719026566, -0.1740163415670395, -0.08952315896749496, 0.025256427004933357, 0.12036951631307602, -0.044322311878204346, 0.11434400081634521, -0.03798634931445122, -0.0008946277666836977, 0.0551130473613739, -0.08121482282876968, -0.12091401219367981, -0.09366606175899506, 0.024188030511140823, 0.04874342679977417, 0.07062821835279465, -0.12095075100660324, -0.0989416241645813, -0.042183902114629745, 0.12152718007564545, -0.09255168586969376, -0.014248261228203773, -0.12553328275680542, 0.059481870383024216, 0.1389842927455902, -0.06352536380290985, 0.03753345087170601, 0.03712745010852814, 0.12064635008573532, 0.029314344748854637, -0.00515587255358696, 0.11027419567108154, -0.08103697746992111, -0.1728755384683609, -0.04343722015619278, 0.1805325448513031, 0.04570537433028221, 0.07732392847537994, -0.008116110228002071, 0.006015656981617212, -0.005607771221548319, -0.06207357347011566, 0.075316421687603, 0.03877919167280197, -0.04523012787103653, 0.04381919279694557, -0.025601880624890327, 0.011734699830412865, -0.09808576852083206, -0.062147848308086395, 0.1569160521030426, 0.27105003595352173, -0.057523176074028015, 0.04276793822646141, 0.05662243813276291, -0.0697113499045372, -0.15242698788642883, 0.015867410227656364, 0.13443177938461304, 0.04472426325082779, -0.0021295922342687845, -0.2301894575357437, 0.048451296985149384, 0.06039836257696152, -0.023920102044939995, 0.041919734328985214, -0.332070529460907, -0.13288825750350952, 0.12203771620988846, 0.09300698339939117, -0.009234046563506126, -0.12917442619800568, -0.06592448055744171, -0.013486913405358791, -0.0652841255068779, 0.02000437304377556, -0.05730669945478439, 0.1378791332244873, 0.024100255221128464, 0.06134844198822975, 0.029286695644259453, -0.04925556480884552, 0.13874489068984985, -0.02450687438249588, 0.0565192848443985, -0.025273410603404045, 0.042275864630937576, 0.010321501642465591, -0.05308543145656586, 0.021485889330506325, -0.08772427588701248, 0.012476211413741112, -0.1311439573764801, -0.03624183312058449, -0.07538916915655136, 0.026720304042100906, -0.02971263974905014, -0.04248955845832825, 0.013772306963801384, 0.03958190232515335, 0.0833941325545311, 0.030693134292960167, 0.10055235773324966, -0.08762094378471375, 0.13959157466888428, 0.09082047641277313, 0.15346889197826385, -0.008165950886905193, -0.08018897473812103, -0.018613439053297043, -0.010185044258832932, 0.04900007322430611, -0.10682114958763123, 0.03978307545185089, 0.13230812549591064, 0.038649093359708786, 0.15655280649662018, 0.05807824060320854, -0.09228954464197159, 0.033559154719114304, 0.055637259036302567, -0.062149785459041595, -0.17487776279449463, -0.034112218767404556, 0.09749976545572281, -0.1538555771112442, 0.012983669526875019, 0.11746355146169662, -0.05581287294626236, -0.0037801440339535475, 0.007113975938409567, 0.021737271919846535, -0.06639096140861511, 0.21717053651809692, 0.020916368812322617, 0.08014395833015442, -0.08230945467948914, 0.0956621989607811, 0.034488026052713394, -0.12155021727085114, 0.05601395666599274, 0.09058420360088348, -0.02456880360841751, -0.01607678271830082, 0.00036503231967799366, 0.0796426311135292, 0.05685087665915489, -0.045969683676958084, -0.13190677762031555, -0.14770302176475525, 0.07096276432275772, 0.10583315789699554, 0.0182807594537735, 0.024146942421793938, -0.05354532226920128, 0.055083103477954865, -0.09598827362060547, 0.09249795228242874, 0.09915313124656677, 0.05272141098976135, -0.14118091762065887, 0.1608027070760727, 0.005433850921690464, 0.0024711089208722115, -0.002948999870568514, -0.028661932796239853, -0.10147132724523544, 0.043838150799274445, -0.10215999186038971, -0.017021413892507553, -0.043021585792303085, -0.0019153989851474762, 0.01722588576376438, -0.059716418385505676, -0.06252805888652802, 0.03883516788482666, -0.12291018664836884, -0.044399093836545944, 0.001671321108005941, 0.07551998645067215, -0.06855794787406921, -0.024986017495393753, 0.05809645727276802, -0.11393197625875473, 0.09641490131616592, 0.058655545115470886, -0.0056442925706505775, 0.02466638572514057, -0.13681749999523163, 0.00948738306760788, 0.025888286530971527, -0.0010614722268655896, 0.004145695827901363, -0.15074889361858368, -0.0313040055334568, -0.026110833510756493, 0.02280534990131855, -0.017797626554965973, 0.0313028022646904, -0.11293879896402359, -0.04724826663732529, -0.005084873177111149, -0.055124443024396896, -0.06379435956478119, 0.042471494525671005, 0.04708718881011009, 0.043230339884757996, 0.14444148540496826, -0.09595032036304474, 0.055127549916505814, -0.21363887190818787, 0.025552431121468544, -0.015092508867383003, -0.07012231647968292, -0.05030934885144234, -0.03628877177834511, 0.08611619472503662, -0.0753888487815857, 0.0632275938987732, -0.055014170706272125, 0.007090565282851458, 0.031314752995967865, -0.12261474877595901, 0.0628931075334549, 0.051218368113040924, 0.24441282451152802, 0.038719311356544495, -0.027323875576257706, 0.08463346213102341, -0.023132018744945526, 0.05117454752326012, 0.15733584761619568, 0.11801053583621979, 0.19105610251426697, 0.0643654465675354, 0.09528498351573944, 0.09278842806816101, -0.06870162487030029, -0.12760640680789948, 0.12096404284238815, -0.008358245715498924, 0.12629830837249756, -0.022281767800450325, 0.2374436855316162, 0.12950627505779266, -0.16494402289390564, 0.05398409068584442, -0.034113332629203796, -0.07917391508817673, -0.10024002194404602, -0.10750265419483185, -0.07223549485206604, -0.16604560613632202, 0.009000622667372227, -0.11060110479593277, 0.048676203936338425, 0.05102110654115677, 0.03714653477072716, 0.020775437355041504, 0.12918594479560852, 0.03322353586554527, -0.0014276126166805625, 0.13445138931274414, -0.025903545320034027, -0.02724776417016983, -0.06137996166944504, -0.11749594658613205, 0.07247207313776016, 0.00009039438009494916, 0.04668908938765526, -0.04326653108000755, -0.09515280276536942, 0.04853559285402298, -0.002138975076377392, -0.1158437579870224, 0.024246957153081894, -0.022963542491197586, 0.0674731507897377, 0.044762060046195984, 0.04496664181351662, -0.024942999705672264, -0.014522532001137733, 0.25327807664871216, -0.10727443546056747, -0.06397458910942078, -0.1472710371017456, 0.2073453962802887, -0.011529981158673763, -0.009009057655930519, -0.002299672458320856, -0.07064061611890793, -0.003047729143872857, 0.16955828666687012, 0.12914958596229553, -0.021568555384874344, -0.013636679388582706, -0.004791394341737032, -0.011631779372692108, -0.06760764122009277, 0.07724322378635406, 0.11586067080497742, -0.001373931998386979, -0.05919564142823219, -0.018382824957370758, -0.032273463904857635, -0.060516104102134705, -0.03747507557272911, 0.07780665904283524, 0.022791508585214615, -0.004169024061411619, -0.029143789783120155, 0.12011144310235977, -0.08774445950984955, -0.09596829861402512, -0.04197310656309128, -0.15302708745002747, -0.17188981175422668, -0.04489266499876976, 0.05222698301076889, 0.05825687572360039, 0.04700983315706253, -0.007403519004583359, 0.0002971900103148073, 0.0825553759932518, 0.007706850301474333, -0.03371766582131386, -0.10701105743646622, 0.09021821618080139, -0.13418906927108765, 0.2006312608718872, -0.04607110843062401, 0.004111467860639095, 0.11821984499692917, 0.059463001787662506, -0.06914454698562622, 0.0491974763572216, 0.06128861382603645, -0.13282588124275208, 0.03636222332715988, 0.2213481366634369, -0.038976311683654785, 0.16866540908813477, 0.031456850469112396, -0.12754513323307037, 0.011182734742760658, -0.08395453542470932, -0.05136789381504059, -0.060046955943107605, -0.014693780802190304, -0.058250460773706436, 0.1304071843624115, 0.20703481137752533, -0.08832249790430069, -0.03125632554292679, -0.05519941449165344, -0.002242364687845111, 0.055979371070861816, 0.08394460380077362, -0.044335510581731796, -0.2843635678291321, -0.003997082356363535, -0.03441530093550682, 0.006235224194824696, -0.232477605342865, -0.09381798654794693, 0.015071406960487366, -0.051249824464321136, -0.042975421994924545, 0.1030147448182106, 0.1193825975060463, 0.04749473184347153, -0.05536973848938942, -0.11563614755868912, -0.018877046182751656, 0.1984037160873413, -0.16036494076251984, -0.052528511732816696 ]
null
null
transformers
# MoECPM Untrained 4x2b ## Model Details ### Model Description A MoE model out of 4 MiniCPM-2B-sft models. Intended to be trained. This version probably does not perform well (if it works at all, lol. I haven't tested it). ## Uses - Training ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
{"language": ["en", "zh"], "license": "apache-2.0", "tags": ["Mixtral", "openbmb/MiniCPM-2B-sft-bf16-llama-format", "MoE", "merge", "mergekit", "moerge", "MiniCPM"], "base_model": ["openbmb/MiniCPM-2B-sft-bf16-llama-format"]}
text-generation
Inv/MoECPM-Untrained-4x2b
[ "transformers", "safetensors", "mixtral", "text-generation", "Mixtral", "openbmb/MiniCPM-2B-sft-bf16-llama-format", "MoE", "merge", "mergekit", "moerge", "MiniCPM", "conversational", "en", "zh", "base_model:openbmb/MiniCPM-2B-sft-bf16-llama-format", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T12:10:34+00:00
[]
[ "en", "zh" ]
TAGS #transformers #safetensors #mixtral #text-generation #Mixtral #openbmb/MiniCPM-2B-sft-bf16-llama-format #MoE #merge #mergekit #moerge #MiniCPM #conversational #en #zh #base_model-openbmb/MiniCPM-2B-sft-bf16-llama-format #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# MoECPM Untrained 4x2b ## Model Details ### Model Description A MoE model out of 4 MiniCPM-2B-sft models. Intended to be trained. This version probably does not perform well (if it works at all, lol. I haven't tested it). ## Uses - Training ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
[ "# MoECPM Untrained 4x2b", "## Model Details", "### Model Description\n\nA MoE model out of 4 MiniCPM-2B-sft models. Intended to be trained. This version probably does not perform well (if it works at all, lol. I haven't tested it).", "## Uses\n\n - Training", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations." ]
[ "TAGS\n#transformers #safetensors #mixtral #text-generation #Mixtral #openbmb/MiniCPM-2B-sft-bf16-llama-format #MoE #merge #mergekit #moerge #MiniCPM #conversational #en #zh #base_model-openbmb/MiniCPM-2B-sft-bf16-llama-format #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# MoECPM Untrained 4x2b", "## Model Details", "### Model Description\n\nA MoE model out of 4 MiniCPM-2B-sft models. Intended to be trained. This version probably does not perform well (if it works at all, lol. I haven't tested it).", "## Uses\n\n - Training", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations." ]
[ 131, 11, 3, 53, 5, 42 ]
[ "passage: TAGS\n#transformers #safetensors #mixtral #text-generation #Mixtral #openbmb/MiniCPM-2B-sft-bf16-llama-format #MoE #merge #mergekit #moerge #MiniCPM #conversational #en #zh #base_model-openbmb/MiniCPM-2B-sft-bf16-llama-format #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# MoECPM Untrained 4x2b## Model Details### Model Description\n\nA MoE model out of 4 MiniCPM-2B-sft models. Intended to be trained. This version probably does not perform well (if it works at all, lol. I haven't tested it).## Uses\n\n - Training### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations." ]
[ -0.08854611963033676, -0.08571867644786835, -0.0034186476841568947, 0.009682427160441875, 0.0671890377998352, -0.013999701477587223, 0.13435028493404388, 0.028683967888355255, 0.0027855061925947666, 0.05159854143857956, 0.05300029739737511, -0.023063521832227707, 0.024408169090747833, 0.15066967904567719, -0.05860418453812599, -0.20362308621406555, 0.07452122867107391, -0.07516609877347946, -0.07396337389945984, 0.08604761213064194, 0.11376060545444489, -0.10049492120742798, 0.06252019852399826, 0.005201385356485844, -0.05981840938329697, 0.006935307290405035, 0.011597748845815659, 0.0081955436617136, 0.0490197129547596, 0.07889765501022339, 0.02950512059032917, 0.05802296847105026, 0.03598363697528839, -0.11483936011791229, 0.049382392317056656, 0.040964990854263306, 0.03176364302635193, 0.054092131555080414, -0.0052979253232479095, 0.004410963971167803, 0.11232511699199677, -0.07383008301258087, 0.05690436810255051, 0.07163728028535843, -0.018100762739777565, -0.07460658997297287, -0.07706142216920853, 0.0801391452550888, 0.20258678495883942, 0.02781710959970951, -0.010950248688459396, 0.2132411003112793, -0.08420965820550919, 0.04732102155685425, 0.17402619123458862, -0.23684775829315186, -0.04052269086241722, 0.17673666775226593, 0.012302346527576447, 0.07067088782787323, -0.03773389756679535, 0.061366092413663864, 0.09017285704612732, -0.007037254050374031, 0.02869618684053421, -0.03851330652832985, 0.14175249636173248, -0.057200998067855835, -0.142917662858963, 0.021324990317225456, 0.24024219810962677, 0.032205358147621155, -0.11315193772315979, -0.21386265754699707, -0.06975620239973068, 0.10920552909374237, -0.03145264461636543, -0.06479133665561676, 0.016159290447831154, -0.015532194636762142, 0.0895785391330719, -0.043869227170944214, -0.04652097821235657, -0.11013686656951904, -0.019525814801454544, 0.1268235445022583, 0.008266007527709007, 0.034175194799900055, -0.039932988584041595, 0.10701388865709305, 0.03650530055165291, -0.09827980399131775, -0.05953260138630867, -0.13226699829101562, 0.00689847394824028, -0.02336416393518448, -0.012984947301447392, -0.038270220160484314, 0.060557443648576736, 0.10563170164823532, 0.004500436130911112, 0.04417373239994049, 0.02197643369436264, 0.05688996613025665, 0.06577561050653458, -0.09359413385391235, -0.056070469319820404, -0.09344489872455597, 0.10796022415161133, 0.09317822009325027, 0.10142641514539719, -0.01976604573428631, -0.04014229029417038, -0.018932541832327843, -0.025412149727344513, 0.1293582171201706, -0.014700972475111485, 0.025888625532388687, -0.02893478237092495, -0.08103252947330475, 0.08420592546463013, -0.12253943830728531, -0.011345135048031807, 0.0035292876418679953, -0.04850907623767853, 0.09754025191068649, 0.047607600688934326, 0.022489825263619423, -0.024271268397569656, 0.0616288036108017, -0.07501965761184692, -0.01842210814356804, -0.09263449162244797, -0.08324937522411346, 0.011106041260063648, 0.09366875141859055, -0.0018735550111159682, -0.13717356324195862, -0.1530803143978119, 0.007590072695165873, 0.00437812739983201, -0.020400531589984894, -0.022088924422860146, 0.007139928638935089, -0.06806966662406921, -0.007029629312455654, -0.031119486317038536, 0.079407699406147, -0.04234969615936279, -0.021031640470027924, -0.00426067691296339, 0.030482424423098564, -0.010425494983792305, -0.02811829000711441, -0.062301501631736755, 0.08903159201145172, -0.10347063839435577, 0.032485343515872955, -0.07182970643043518, 0.022436680272221565, -0.11463452130556107, -0.055999547243118286, -0.06605099141597748, 0.009437143802642822, 0.02080058865249157, 0.2221667319536209, -0.13723555207252502, -0.001255640760064125, 0.22506080567836761, -0.16222673654556274, -0.10994365811347961, 0.14671233296394348, -0.022655071690678596, 0.03888959810137749, 0.08632589131593704, 0.0028439818415790796, 0.11981002986431122, -0.13615673780441284, 0.024847812950611115, 0.054608896374702454, 0.03721288964152336, 0.06814402341842651, 0.09754538536071777, 0.00431657861918211, -0.10784770548343658, 0.012911910191178322, -0.007324958685785532, -0.015999196097254753, -0.04332096129655838, -0.026958046481013298, -0.06380251795053482, -0.022456198930740356, 0.024300141260027885, -0.010968392714858055, -0.025119394063949585, -0.07332494854927063, -0.05511830747127533, 0.05503790080547333, 0.14549586176872253, -0.030671291053295135, -0.028099609538912773, -0.12807859480381012, 0.05343032255768776, -0.06005100905895233, 0.022551415488123894, -0.09335378557443619, -0.05254845693707466, -0.033828962594270706, -0.06264019757509232, -0.002654528245329857, 0.18402500450611115, 0.036041680723428726, 0.09512895345687866, -0.02417081780731678, 0.0002767416008282453, 0.06028260663151741, 0.005280268378555775, -0.024807775393128395, -0.21093183755874634, -0.013039663434028625, -0.05358164757490158, 0.15249232947826385, -0.2505994737148285, 0.040918443351984024, 0.06640123575925827, 0.10869278013706207, 0.02714918926358223, 0.010667279362678528, 0.0563756488263607, -0.0472981259226799, 0.008074283599853516, -0.06576226651668549, 0.03667968511581421, -0.005495505407452583, -0.07338262349367142, 0.06263934820890427, -0.21202920377254486, 0.1325298547744751, 0.1332339644432068, 0.07436703145503998, -0.0344838872551918, -0.09904490411281586, -0.04127347841858864, -0.007774121128022671, -0.02617466077208519, -0.06166259944438934, 0.05109823867678642, 0.03491329774260521, 0.1173405647277832, -0.10556699335575104, -0.06967663764953613, 0.01323008444160223, -0.05725495517253876, 0.006622503977268934, 0.01718749664723873, 0.05009607598185539, -0.20649805665016174, 0.03502031788229942, 0.0974593535065651, 0.02991381846368313, 0.12232901155948639, 0.023379987105727196, -0.042651187628507614, -0.018391015008091927, 0.035221029072999954, -0.0028911668341606855, 0.13034974038600922, 0.0018119034357368946, 0.04659212753176689, 0.06474199891090393, 0.06286169588565826, 0.03551848232746124, -0.06648439168930054, 0.00381206045858562, 0.015964539721608162, -0.03684832155704498, 0.029109880328178406, 0.05378025770187378, -0.023762967437505722, 0.12007570266723633, 0.004956004675477743, 0.06822065263986588, -0.004977135919034481, -0.08110371977090836, -0.14057248830795288, 0.1666175127029419, -0.07219522446393967, -0.178655207157135, -0.1108742356300354, 0.038376763463020325, -0.06267297267913818, -0.016567513346672058, 0.002517491113394499, -0.02691304311156273, -0.12370291352272034, -0.12154689431190491, 0.09739712625741959, -0.0046557290479540825, -0.0416325218975544, 0.01696503534913063, 0.02150614932179451, 0.07327473163604736, -0.15964548289775848, -0.01198774017393589, 0.025341946631669998, -0.0899367704987526, -0.013790945522487164, -0.013337529264390469, 0.035169653594493866, 0.1328822374343872, 0.01628708466887474, 0.010400054976344109, 0.01292898878455162, 0.18240706622600555, -0.028894459828734398, -0.009502477012574673, 0.2506945729255676, -0.027262238785624504, 0.01262995507568121, 0.13353733718395233, 0.03273017704486847, -0.11092668771743774, 0.049226898699998856, 0.012309897691011429, -0.04932146891951561, -0.1481664627790451, -0.12178421765565872, -0.054920826107263565, -0.0018617419991642237, 0.002115177921950817, 0.04489000886678696, -0.007046875543892384, 0.06393679976463318, -0.14710822701454163, -0.0138776283711195, 0.06033095717430115, 0.08836975693702698, 0.07212892174720764, -0.01922302320599556, 0.10892945528030396, -0.06666495651006699, 0.0006295890198089182, 0.10029447078704834, -0.08279789984226227, 0.15025343000888824, 0.0031179694924503565, 0.06713447719812393, 0.12066422402858734, 0.007963045500218868, 0.06596171110868454, 0.08171909302473068, -0.018236227333545685, -0.04340815544128418, -0.03198443725705147, -0.11105319857597351, -0.0750778391957283, 0.04583791643381119, -0.0502939447760582, 0.15835975110530853, -0.08194759488105774, 0.05447829142212868, 0.05030800774693489, 0.10287472605705261, 0.042422082275152206, -0.22940026223659515, -0.10205210000276566, 0.08636096864938736, -0.022975312545895576, -0.03719412162899971, 0.02230568788945675, 0.174860879778862, -0.1346943974494934, 0.05122232064604759, -0.03894590213894844, 0.096035435795784, 0.00084717757999897, 0.005707130767405033, -0.06138484925031662, 0.09157218039035797, -0.01527419313788414, 0.07344361394643784, -0.17214597761631012, 0.07397788017988205, 0.026260165497660637, 0.09444870799779892, -0.05307247117161751, 0.01901237666606903, 0.0827040821313858, 0.18148477375507355, 0.10855668783187866, 0.02771706134080887, -0.020577479153871536, -0.04666918143630028, -0.09772414714097977, 0.04023706540465355, 0.038474105298519135, -0.016042210161685944, 0.08670041710138321, -0.08739083260297775, -0.016507508233189583, 0.01039261743426323, -0.01309203915297985, -0.22232691943645477, -0.11570718884468079, 0.025533264502882957, -0.0029289727099239826, -0.05836385115981102, -0.12643727660179138, -0.04267838969826698, 0.06918714940547943, 0.2244519591331482, -0.19549164175987244, -0.09567626565694809, -0.11514163762331009, 0.008053439669311047, 0.03917667269706726, -0.01570247672498226, 0.005998935084789991, -0.023481249809265137, 0.208729088306427, -0.005994000937789679, -0.12190904468297958, 0.026652486994862556, -0.15527836978435516, -0.15616492927074432, -0.036601997911930084, 0.17311881482601166, 0.007007946260273457, 0.05520186573266983, 0.04882251098752022, -0.006417396944016218, 0.0357685349881649, -0.17100073397159576, -0.08575563877820969, 0.20245932042598724, 0.12742190062999725, 0.06937498599290848, -0.04086883366107941, -0.04166184738278389, 0.03851030766963959, 0.0172603577375412, 0.08638349175453186, 0.3009907901287079, -0.012764106504619122, 0.10423096269369125, 0.06396458297967911, -0.019368819892406464, -0.2060685008764267, 0.023869525641202927, -0.020457221195101738, 0.041254665702581406, 0.01816059648990631, -0.0327107273042202, 0.10876893997192383, -0.0071472167037427425, -0.030094947665929794, 0.16708478331565857, -0.31717449426651, -0.11757711321115494, 0.048849694430828094, 0.11741743981838226, 0.20467664301395416, -0.07702827453613281, -0.05149237439036369, -0.1010497659444809, -0.15473513305187225, 0.12567758560180664, -0.2272707223892212, 0.030783195048570633, -0.01738494634628296, -0.004093862604349852, 0.03513879328966141, -0.047724924981594086, 0.19473087787628174, -0.10997049510478973, 0.06772109866142273, -0.10302525758743286, 0.05508095771074295, 0.05341525375843048, -0.052839286625385284, 0.11375046521425247, -0.09440279752016068, 0.049478769302368164, -0.12099239230155945, -0.07944563776254654, 0.01716284640133381, 0.06202718988060951, -0.0606076605618, -0.05553845688700676, -0.0462460033595562, 0.07032374292612076, 0.033106718212366104, 0.014983470551669598, -0.02655131369829178, -0.10302591323852539, -0.012858263216912746, 0.12546730041503906, 0.2330925464630127, -0.06921140849590302, 0.0126110864803195, 0.00911027193069458, 0.007874695584177971, 0.10837461054325104, -0.13329021632671356, 0.058293867856264114, 0.02872118167579174, 0.02246280014514923, 0.04476350173354149, 0.018125144764780998, -0.09221182018518448, 0.0017316973535344005, 0.05951562896370888, -0.1695224791765213, -0.10056540369987488, -0.021565457805991173, 0.16539183259010315, -0.0616622269153595, -0.003277321346104145, 0.15642109513282776, -0.07283855974674225, 0.024830076843500137, -0.01222402136772871, 0.007466169074177742, -0.04884769394993782, 0.08485032618045807, 0.03699985519051552, 0.08923225849866867, -0.07404415309429169, 0.01891733519732952, -0.03579697385430336, -0.02019794099032879, 0.004812134429812431, 0.004454813431948423, -0.12017466872930527, -0.06864643841981888, -0.023812860250473022, 0.1159646138548851, 0.13074876368045807, -0.10600274801254272, -0.06160310283303261, -0.13930927217006683, -0.028717627748847008, 0.04379187524318695, 0.08213771879673004, -0.0046288068406283855, 0.013509630225598812, -0.004774360917508602, -0.09607941657304764, 0.06650048494338989, 0.0176200233399868, 0.021266689524054527, -0.12157358229160309, 0.07158899307250977, 0.0235938411206007, 0.0064421119168400764, -0.07384451478719711, 0.0004552871105261147, -0.0587359219789505, -0.02671334706246853, -0.2553081810474396, 0.024262171238660812, -0.07868247479200363, -0.026009486988186836, -0.014613164588809013, -0.008170864544808865, -0.08149567991495132, 0.044972676783800125, -0.02207903377711773, 0.007874706760048866, -0.03297746926546097, 0.030447550117969513, -0.08101030439138412, 0.002890398260205984, 0.03267955034971237, -0.09498840570449829, 0.060863953083753586, 0.022170886397361755, -0.02936880849301815, 0.023530155420303345, -0.07758283615112305, -0.01299557089805603, 0.03503892198204994, 0.03802298754453659, 0.0016338912537321448, -0.11674617230892181, 0.03520838916301727, 0.013997453264892101, 0.009796425700187683, 0.0024900392163544893, 0.07829820364713669, -0.07510516047477722, -0.03280021250247955, -0.060870666056871414, -0.005679934751242399, -0.05476434901356697, -0.005385514348745346, 0.02059459313750267, 0.0557418093085289, 0.16310207545757294, -0.05810634791851044, 0.015347064472734928, -0.14760735630989075, -0.03886132687330246, 0.004721108824014664, -0.007478783372789621, 0.003220020094886422, 0.002162478631362319, 0.035877298563718796, -0.03650425374507904, 0.20860788226127625, -0.01792163960635662, -0.025105014443397522, 0.07045699656009674, -0.08083761483430862, 0.11008104681968689, 0.040840454399585724, 0.15873906016349792, 0.029698742553591728, 0.06915341317653656, -0.053269486874341965, 0.004937571007758379, 0.0769805759191513, -0.13983607292175293, 0.17967799305915833, 0.15034028887748718, -0.07054334133863449, 0.07114452123641968, -0.01739497110247612, -0.07779429107904434, 0.007983584888279438, -0.002240993781015277, 0.0018915710970759392, 0.010249567218124866, -0.01704765297472477, 0.06083672121167183, 0.24330371618270874, -0.12805771827697754, -0.0031527027022093534, -0.047433752566576004, -0.032518282532691956, -0.15355713665485382, -0.14540715515613556, -0.06259030848741531, -0.0935693085193634, -0.009229819290339947, -0.11222071945667267, -0.035721950232982635, 0.05828407406806946, 0.04468885809183121, -0.004782648757100105, 0.15330712497234344, -0.15605230629444122, -0.027364064007997513, 0.02423357032239437, -0.008616320788860321, 0.00040064938366413116, 0.04524032026529312, -0.033544398844242096, 0.07861413806676865, -0.051037080585956573, 0.021692968904972076, 0.04289333522319794, 0.01321911159902811, -0.007462345529347658, -0.027521973475813866, -0.08449601382017136, -0.023355742916464806, 0.030423620715737343, 0.051644373685121536, 0.11887365579605103, 0.06388111412525177, -0.04709836095571518, -0.031922630965709686, 0.16100208461284637, -0.03832229599356651, -0.12080701440572739, -0.09772993624210358, 0.1127091720700264, -0.01500408910214901, 0.07658965140581131, 0.029274508357048035, -0.05681207403540611, 0.03259541839361191, 0.11992661654949188, 0.21916474401950836, -0.14046788215637207, 0.01658531464636326, -0.015353160910308361, 0.002322223735973239, -0.022832907736301422, 0.14990617334842682, 0.0413387231528759, 0.18268650770187378, -0.019977660849690437, 0.051478177309036255, -0.02389742061495781, -0.05326669290661812, -0.07479142397642136, 0.07897555828094482, 0.034670140594244, -0.006285478826612234, -0.0007505977991968393, 0.07785049825906754, -0.011089521460235119, -0.05417536199092865, 0.028660159558057785, -0.04382777959108353, -0.11883368343114853, -0.006238476373255253, -0.0627656951546669, 0.00009331112960353494, 0.043104398995637894, -0.06677786260843277, -0.026458997279405594, 0.10511918365955353, 0.003977879881858826, -0.041407011449337006, -0.062181126326322556, 0.09558796882629395, -0.0031813839450478554, 0.17303605377674103, 0.02058481238782406, 0.13137945532798767, 0.08146518468856812, -0.09633270651102066, -0.09622011333703995, 0.001543231774121523, 0.04011664539575577, -0.06841030716896057, 0.02112565189599991, 0.14519014954566956, -0.020271895453333855, 0.05566398799419403, 0.05771457403898239, -0.11139975488185883, -0.015612109564244747, -0.03850330039858818, -0.07384934276342392, -0.07247497141361237, 0.15906938910484314, -0.07034984230995178, 0.14338022470474243, 0.0764884501695633, -0.06882856041193008, 0.01376942079514265, -0.03677383437752724, 0.14078140258789062, 0.03556789085268974, 0.05484626442193985, 0.05655165761709213, -0.17036324739456177, 0.00640525110065937, 0.03427319601178169, 0.044139619916677475, -0.31683990359306335, -0.07629711925983429, -0.04414254426956177, 0.016352981328964233, -0.04803182929754257, 0.08339773863554001, 0.05514060705900192, 0.03435729444026947, -0.04800756275653839, -0.206887349486351, -0.02136719971895218, 0.07770466059446335, -0.09428125619888306, -0.07537634670734406 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text-generation
aidonuts/marigold-005-ep4
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T12:12:22+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 60, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.04654794931411743, 0.16618601977825165, -0.005445904564112425, 0.01853804849088192, 0.0981811136007309, 0.011998992413282394, 0.06433123350143433, 0.11398410052061081, -0.0230073444545269, 0.11406639218330383, 0.03047988750040531, 0.10172267258167267, 0.11317981779575348, 0.14841650426387787, -0.002152352826669812, -0.22403094172477722, 0.050844956189394, -0.12105348706245422, -0.033293843269348145, 0.11749980598688126, 0.1483822613954544, -0.09928343445062637, 0.07274559140205383, -0.029687678441405296, -0.012143402360379696, -0.030057786032557487, -0.05890674889087677, -0.046214159578084946, 0.04651786759495735, 0.06640566885471344, 0.06770290434360504, 0.0071083661168813705, 0.09012923389673233, -0.2696533799171448, 0.018959321081638336, 0.07145345956087112, -0.002759667346253991, 0.06957992166280746, 0.06404146552085876, -0.07107418030500412, 0.10337356477975845, -0.05106033384799957, 0.14650006592273712, 0.08365883678197861, -0.09081148356199265, -0.1895141303539276, -0.08866965025663376, 0.09882009029388428, 0.17572562396526337, 0.04925641790032387, -0.02320658043026924, 0.09761467576026917, -0.08769196271896362, 0.015438909642398357, 0.04981724172830582, -0.07620415836572647, -0.05378096550703049, 0.05986575037240982, 0.07907199114561081, 0.06627275794744492, -0.12434766441583633, -0.02885502204298973, 0.005009706597775221, 0.010980482213199139, 0.0769270583987236, 0.01728810742497444, 0.146672785282135, 0.0338633768260479, -0.12615777552127838, -0.04880760237574577, 0.09869225323200226, 0.03395522013306618, -0.04422314465045929, -0.24749068915843964, -0.03152675926685333, -0.030810698866844177, -0.029386121779680252, -0.03716538846492767, 0.04340358078479767, -0.007673026993870735, 0.08638741075992584, -0.0060646249912679195, -0.07403432577848434, -0.03937075287103653, 0.06169692054390907, 0.0672287791967392, 0.02999979443848133, -0.013745363801717758, 0.010938193649053574, 0.11620724946260452, 0.1095694974064827, -0.12054188549518585, -0.05555335059762001, -0.06393084675073624, -0.08656639605760574, -0.040790557861328125, 0.034162238240242004, 0.03456587344408035, 0.05349370837211609, 0.25305667519569397, 0.015654386952519417, 0.059652652591466904, 0.034477248787879944, 0.007892133668065071, 0.05848940089344978, 0.11044429242610931, -0.06018859148025513, -0.10444226115942001, -0.02648012898862362, 0.08843598514795303, 0.008199662901461124, -0.03287925571203232, -0.05088530853390694, 0.06019928678870201, 0.01946467161178589, 0.11926145106554031, 0.09061790257692337, 0.010536285117268562, -0.07121123373508453, -0.061038948595523834, 0.1891259253025055, -0.16544590890407562, 0.04322727024555206, 0.035097137093544006, -0.03903156518936157, 0.00019933005387429148, 0.013914269395172596, 0.016625655815005302, -0.025983380153775215, 0.09017423540353775, -0.054113563150167465, -0.04145489260554314, -0.11186197400093079, -0.03383193537592888, 0.033762916922569275, 0.008953776210546494, -0.035059962421655655, -0.033713940531015396, -0.08351044356822968, -0.07577689737081528, 0.09320491552352905, -0.07346344739198685, -0.04878907650709152, -0.01804324984550476, -0.07530532777309418, 0.022395428270101547, 0.019394835457205772, 0.07707412540912628, -0.02362251654267311, 0.04399976506829262, -0.05189276114106178, 0.05863580107688904, 0.11207318305969238, 0.03570080175995827, -0.05736649036407471, 0.06062258034944534, -0.23834340274333954, 0.09552820026874542, -0.07409077137708664, 0.05591456592082977, -0.153293639421463, -0.024439791217446327, 0.04788333550095558, 0.008784620091319084, -0.009650949388742447, 0.13416339457035065, -0.21702027320861816, -0.02536402828991413, 0.1717337965965271, -0.10057014971971512, -0.07069246470928192, 0.05619903281331062, -0.04835370555520058, 0.10988964140415192, 0.03825836628675461, -0.025690359994769096, 0.06171267107129097, -0.1267417073249817, 0.003717758459970355, -0.05005312338471413, -0.017048977315425873, 0.1548657864332199, 0.07182947546243668, -0.07217690348625183, 0.07399354875087738, 0.025708531960844994, -0.0246540866792202, -0.04625825211405754, -0.015164627693593502, -0.10536660254001617, 0.014689887873828411, -0.06369215250015259, 0.014470234513282776, -0.020807426422834396, -0.09071163833141327, -0.027962757274508476, -0.17504668235778809, -0.03014434315264225, 0.08651752024888992, -0.008693269453942776, -0.01803150773048401, -0.1178668737411499, 0.009341353550553322, 0.04177580401301384, 0.0061247628182172775, -0.13462838530540466, -0.04812471568584442, 0.02780051715672016, -0.1600649207830429, 0.034652888774871826, -0.05392369255423546, 0.04932025074958801, 0.025790516287088394, -0.028889117762446404, -0.026493212208151817, 0.021633783355355263, 0.005992184858769178, -0.011999987065792084, -0.24343903362751007, -0.028118690475821495, -0.024888472631573677, 0.1682123839855194, -0.20917098224163055, 0.03546025976538658, 0.07867541164159775, 0.15366052091121674, 0.011240328662097454, -0.04177491366863251, 0.005974748637527227, -0.06935794651508331, -0.02736494317650795, -0.05875484645366669, -0.0047869328409433365, -0.03310677409172058, -0.04545191675424576, 0.04568447172641754, -0.16510973870754242, -0.032636504620313644, 0.09776268899440765, 0.06289951503276825, -0.13922683894634247, -0.020621931180357933, -0.03630133345723152, -0.049253206700086594, -0.04911839962005615, -0.0605199858546257, 0.10893940925598145, 0.05891856551170349, 0.04574795812368393, -0.05928509309887886, -0.07568105310201645, -0.001827909960411489, -0.013898161239922047, -0.017864689230918884, 0.09759635478258133, 0.0751434788107872, -0.13251115381717682, 0.09224759042263031, 0.09603385627269745, 0.07919023185968399, 0.09113933145999908, -0.02355697751045227, -0.08261934667825699, -0.045987509191036224, 0.031442027539014816, 0.020124373957514763, 0.13039541244506836, -0.024294709786772728, 0.04352088272571564, 0.042134687304496765, -0.019369594752788544, 0.014752166345715523, -0.08687400817871094, 0.033972494304180145, 0.028472330421209335, -0.016721390187740326, 0.050190530717372894, -0.03876714035868645, 0.02440318465232849, 0.08830609917640686, 0.045322712510824203, 0.03507532551884651, 0.015493292361497879, -0.05206458270549774, -0.1083620935678482, 0.16405931115150452, -0.12714070081710815, -0.22483378648757935, -0.13936103880405426, 0.0037376401014626026, 0.035628627985715866, -0.015835661441087723, 0.002417160663753748, -0.059374887496232986, -0.12220635265111923, -0.08858037739992142, 0.015140829607844353, 0.04942670464515686, -0.09028962254524231, -0.06437795609235764, 0.058117836713790894, 0.03889724239706993, -0.14560972154140472, 0.017612040042877197, 0.04854894429445267, -0.09789852797985077, -0.006774199660867453, 0.08094939589500427, 0.0698540136218071, 0.1770169734954834, 0.017703235149383545, -0.021850809454917908, 0.032354529947042465, 0.20614571869373322, -0.13538233935832977, 0.11083246022462845, 0.13607586920261383, -0.09041404724121094, 0.08072979003190994, 0.19951270520687103, 0.03932560607790947, -0.10153959691524506, 0.031980328261852264, 0.02283124253153801, -0.0284719280898571, -0.24526868760585785, -0.07212468236684799, -0.004402178805321455, -0.058010730892419815, 0.07660572230815887, 0.09286724030971527, 0.08215958625078201, 0.012304253876209259, -0.09310996532440186, -0.08154371380805969, 0.05942574888467789, 0.10367169976234436, 0.024584239348769188, -0.010839897207915783, 0.08998730033636093, -0.034100502729415894, 0.019626356661319733, 0.0853661298751831, 0.005239574704319239, 0.17840281128883362, 0.05159219726920128, 0.18830420076847076, 0.07925192266702652, 0.07219027727842331, 0.009912233799695969, 0.013080619275569916, 0.018877580761909485, 0.03300119563937187, -0.002769160782918334, -0.08440786600112915, -0.02248465269804001, 0.11566436290740967, 0.06668911874294281, 0.010815348476171494, 0.015172341838479042, -0.04104290530085564, 0.07965951412916183, 0.1831512451171875, -0.007656289264559746, -0.1783534437417984, -0.057547420263290405, 0.07553383708000183, -0.09879875183105469, -0.09854305535554886, -0.013454320840537548, 0.03072015568614006, -0.17046253383159637, 0.023390959948301315, -0.02239842526614666, 0.1106182336807251, -0.14194999635219574, -0.020490378141403198, 0.07218493521213531, 0.07199500501155853, 0.004729843698441982, 0.05758659541606903, -0.16417601704597473, 0.10671813786029816, 0.008950476534664631, 0.06779605895280838, -0.09610627591609955, 0.1008887067437172, -0.004196076653897762, -0.02063460275530815, 0.1393408179283142, 0.002700034761801362, -0.06884108483791351, -0.0763031542301178, -0.08754398673772812, -0.009632662869989872, 0.12754282355308533, -0.1419651061296463, 0.08767123520374298, -0.037212442606687546, -0.0424150750041008, -0.0017086371080949903, -0.10206665843725204, -0.11638247221708298, -0.18888559937477112, 0.06001543253660202, -0.13492922484874725, 0.03152317553758621, -0.10799519717693329, -0.032371897250413895, -0.030304040759801865, 0.19337286055088043, -0.23447458446025848, -0.07199826091527939, -0.1475764364004135, -0.10233612358570099, 0.1443224400281906, -0.0501345656812191, 0.08485390990972519, -0.007241467013955116, 0.16846685111522675, 0.019060896709561348, -0.02531743235886097, 0.0971490666270256, -0.09173708409070969, -0.19302815198898315, -0.07869284600019455, 0.15662524104118347, 0.13260218501091003, 0.031680017709732056, -0.002461588243022561, 0.036563750356435776, -0.015421539545059204, -0.11935004591941833, 0.015969349071383476, 0.1787186712026596, 0.06237189099192619, 0.02331034652888775, -0.027346095070242882, -0.11273157596588135, -0.06900003552436829, -0.028530338779091835, 0.03054865077137947, 0.17762407660484314, -0.07057618349790573, 0.18207968771457672, 0.14163152873516083, -0.05922834202647209, -0.20400173962116241, 0.010538800619542599, 0.03055560030043125, 0.0009220078936778009, 0.02591954916715622, -0.20123432576656342, 0.08688826113939285, 0.004683020059019327, -0.05110127478837967, 0.13194532692432404, -0.17217805981636047, -0.14451217651367188, 0.0765485092997551, 0.038384392857551575, -0.19559739530086517, -0.12913893163204193, -0.09174312651157379, -0.045869920402765274, -0.18591414391994476, 0.09569250047206879, 0.0305706188082695, 0.010893458500504494, 0.03030681423842907, 0.029179483652114868, 0.019487828016281128, -0.0418255440890789, 0.18391458690166473, -0.024792250245809555, 0.026594700291752815, -0.08539514988660812, -0.06927408277988434, 0.03743394836783409, -0.052842434495687485, 0.07349982857704163, -0.023486759513616562, 0.007861839607357979, -0.10348054021596909, -0.042148489505052567, -0.03735732287168503, 0.015448716469109058, -0.09657872468233109, -0.08514349907636642, -0.045032672584056854, 0.09675803780555725, 0.09690850973129272, -0.033646680414676666, -0.028050623834133148, -0.07533035427331924, 0.04412057250738144, 0.19926515221595764, 0.1785389482975006, 0.042153384536504745, -0.08034496754407883, -0.004150947090238333, -0.010121207684278488, 0.04310847446322441, -0.20463712513446808, 0.06283636391162872, 0.05450061708688736, 0.01973269321024418, 0.11436162889003754, -0.019565396010875702, -0.15359151363372803, -0.07263088971376419, 0.06303015351295471, -0.060181066393852234, -0.19620554149150848, 0.00867035984992981, 0.060603946447372437, -0.16371412575244904, -0.04535605385899544, 0.04643881320953369, -0.005620351992547512, -0.038163937628269196, 0.021896906197071075, 0.09194854646921158, 0.0026654244866222143, 0.07427921891212463, 0.05387866869568825, 0.0827430784702301, -0.10537070035934448, 0.08090532571077347, 0.08839722722768784, -0.08452684432268143, 0.023530138656497, 0.10478579998016357, -0.059433579444885254, -0.03440561518073082, 0.020135708153247833, 0.08153781294822693, 0.01775863952934742, -0.040019966661930084, 0.013229827396571636, -0.10452935844659805, 0.05954122915863991, 0.08839859813451767, 0.032507482916116714, 0.016702456399798393, 0.03425082191824913, 0.04607953503727913, -0.07238735258579254, 0.12142276018857956, 0.031868141144514084, 0.017129309475421906, -0.036505792289972305, -0.040896978229284286, 0.019542274996638298, -0.03214648738503456, -0.005015232600271702, -0.03023446537554264, -0.07695909589529037, -0.014793801121413708, -0.1626158058643341, -0.011131818406283855, -0.05648450180888176, 0.010329355485737324, 0.03204665705561638, -0.032609567046165466, 0.008124498650431633, 0.009250079281628132, -0.07695289701223373, -0.0663459524512291, -0.020460480824112892, 0.09540658444166183, -0.16213038563728333, 0.022481130436062813, 0.08244425803422928, -0.12187694013118744, 0.09281346201896667, 0.016204802319407463, -0.006236857734620571, 0.025038830935955048, -0.1475188434123993, 0.034843120723962784, -0.03386561945080757, 0.010836300440132618, 0.04373383894562721, -0.21569781005382538, -0.00004886732858722098, -0.033673107624053955, -0.06639216095209122, -0.009451326914131641, -0.03672455996274948, -0.11508306115865707, 0.1058407872915268, 0.007236586883664131, -0.08753558248281479, -0.03186136856675148, 0.029325377196073532, 0.0838974118232727, -0.021959776058793068, 0.15145497024059296, -0.008370938710868359, 0.07429654151201248, -0.16209737956523895, -0.018623165786266327, -0.006028574425727129, 0.022658247500658035, -0.01664556935429573, -0.01111356820911169, 0.044031109660863876, -0.022746501490473747, 0.17925859987735748, -0.030318550765514374, 0.02272745408117771, 0.06815794110298157, 0.019072026014328003, -0.030184008181095123, 0.10406795144081116, 0.04094860330224037, 0.02014910988509655, 0.018591465428471565, 0.003289656015112996, -0.04647882282733917, -0.03173251822590828, -0.19407226145267487, 0.07288651913404465, 0.15608493983745575, 0.09729263186454773, -0.016707008704543114, 0.07954329252243042, -0.10199416428804398, -0.1109243705868721, 0.12477338314056396, -0.04797708988189697, -0.002418199321255088, -0.07150927931070328, 0.13247236609458923, 0.1437523066997528, -0.1859612911939621, 0.07269313186407089, -0.0699717253446579, -0.04708027467131615, -0.10980689525604248, -0.19441905617713928, -0.05561789125204086, -0.049456022679805756, -0.016053348779678345, -0.04698808491230011, 0.07504211366176605, 0.054538097232580185, 0.006766852922737598, -0.0023397188633680344, 0.06506035476922989, -0.031050674617290497, -0.0037882844917476177, 0.032597362995147705, 0.06591679900884628, 0.012734474614262581, -0.030802709981799126, 0.016619903966784477, -0.013545602560043335, 0.045626189559698105, 0.06578011065721512, 0.04976864159107208, -0.02938537672162056, 0.014603170566260815, -0.038539156317710876, -0.10249634087085724, 0.043612558394670486, -0.024421939626336098, -0.0789753645658493, 0.15477414429187775, 0.023680059239268303, 0.007779473438858986, -0.020137663930654526, 0.23901568353176117, -0.0738423764705658, -0.0964353010058403, -0.14737580716609955, 0.10557299107313156, -0.038081806153059006, 0.05800395458936691, 0.04625935107469559, -0.10226529091596603, 0.018044332042336464, 0.1338089406490326, 0.16182038187980652, -0.039008259773254395, 0.020095856860280037, 0.031135575845837593, 0.00566398398950696, -0.03622615709900856, 0.04847532883286476, 0.06906453520059586, 0.16569648683071136, -0.04632584750652313, 0.09100406616926193, 0.0019041687482967973, -0.09579581767320633, -0.038361791521310806, 0.11069868505001068, -0.016052277758717537, 0.019335128366947174, -0.05818064883351326, 0.11742528527975082, -0.06386786699295044, -0.23783175647258759, 0.06453443318605423, -0.0684293657541275, -0.13765870034694672, -0.02378307841718197, 0.08207765966653824, -0.012955902144312859, 0.027587108314037323, 0.0730307325720787, -0.07240920513868332, 0.201939657330513, 0.03798431158065796, -0.05499868467450142, -0.055047210305929184, 0.0805421993136406, -0.10008571296930313, 0.2739645540714264, 0.01557221356779337, 0.04601577669382095, 0.10384146869182587, -0.009341772645711899, -0.13838784396648407, 0.019836371764540672, 0.09581108391284943, -0.10502193123102188, 0.04196618124842644, 0.19815568625926971, -0.0014755994779989123, 0.12389086186885834, 0.07657600939273834, -0.07551808655261993, 0.0478031262755394, -0.08054235577583313, -0.06760486960411072, -0.09260394424200058, 0.09703279286623001, -0.07772123068571091, 0.14251399040222168, 0.13876807689666748, -0.05074559152126312, 0.012724342755973339, -0.031311117112636566, 0.044293127954006195, -0.00010600237874314189, 0.10321761667728424, 0.004272161517292261, -0.1832672357559204, 0.024692710489034653, 0.005650998093187809, 0.10749758034944534, -0.16033467650413513, -0.09566054493188858, 0.042343202978372574, 0.003505636239424348, -0.0672195628285408, 0.1290110945701599, 0.05665452033281326, 0.04342988133430481, -0.03997718170285225, -0.03521440550684929, -0.0060732318088412285, 0.13561366498470306, -0.10713256150484085, 0.0009933578548952937 ]
null
null
transformers
dict( name="igenius-llama-2b", # official hf_config=dict(org="Igenius", name="igenius-llama-2b"), block_size=2048, vocab_size=32000, padding_multiple=64, n_layer=22, n_head=64, n_embd=2560, rotary_percentage=1.0, parallel_residual=False, bias=False, _norm_class="RMSNorm", norm_eps=1e-5, _mlp_class="LLaMAMLP", intermediate_size=8960, n_query_groups=8, ), Trained on wikipedia and editorial corpus for about 1 epoch
{}
text-generation
iGenius-AI-Team/Italia-2B-ckpt-30B-wikiEditorial
[ "transformers", "safetensors", "llama", "text-generation", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T12:14:46+00:00
[]
[]
TAGS #transformers #safetensors #llama #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
dict( name="igenius-llama-2b", # official hf_config=dict(org="Igenius", name="igenius-llama-2b"), block_size=2048, vocab_size=32000, padding_multiple=64, n_layer=22, n_head=64, n_embd=2560, rotary_percentage=1.0, parallel_residual=False, bias=False, _norm_class="RMSNorm", norm_eps=1e-5, _mlp_class="LLaMAMLP", intermediate_size=8960, n_query_groups=8, ), Trained on wikipedia and editorial corpus for about 1 epoch
[ "# official\n hf_config=dict(org=\"Igenius\", name=\"igenius-llama-2b\"),\n block_size=2048,\n vocab_size=32000,\n padding_multiple=64,\n n_layer=22,\n n_head=64,\n n_embd=2560,\n rotary_percentage=1.0,\n parallel_residual=False,\n bias=False,\n _norm_class=\"RMSNorm\",\n norm_eps=1e-5,\n _mlp_class=\"LLaMAMLP\",\n intermediate_size=8960,\n n_query_groups=8,\n ),\n\n Trained on wikipedia and editorial corpus for about 1 epoch" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# official\n hf_config=dict(org=\"Igenius\", name=\"igenius-llama-2b\"),\n block_size=2048,\n vocab_size=32000,\n padding_multiple=64,\n n_layer=22,\n n_head=64,\n n_embd=2560,\n rotary_percentage=1.0,\n parallel_residual=False,\n bias=False,\n _norm_class=\"RMSNorm\",\n norm_eps=1e-5,\n _mlp_class=\"LLaMAMLP\",\n intermediate_size=8960,\n n_query_groups=8,\n ),\n\n Trained on wikipedia and editorial corpus for about 1 epoch" ]
[ 47, 153 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# official\n hf_config=dict(org=\"Igenius\", name=\"igenius-llama-2b\"),\n block_size=2048,\n vocab_size=32000,\n padding_multiple=64,\n n_layer=22,\n n_head=64,\n n_embd=2560,\n rotary_percentage=1.0,\n parallel_residual=False,\n bias=False,\n _norm_class=\"RMSNorm\",\n norm_eps=1e-5,\n _mlp_class=\"LLaMAMLP\",\n intermediate_size=8960,\n n_query_groups=8,\n ),\n\n Trained on wikipedia and editorial corpus for about 1 epoch" ]
[ -0.07252874225378036, -0.02175043523311615, -0.004738959483802319, 0.05557278171181679, 0.05883242189884186, 0.0396529957652092, 0.16303539276123047, 0.1055101752281189, 0.10876207798719406, 0.023670950904488564, 0.1252809315919876, 0.05201409012079239, 0.019620994105935097, -0.03874560818076134, -0.11741077899932861, -0.16234508156776428, 0.03618019446730614, 0.037073057144880295, 0.013950972817838192, 0.022378847002983093, 0.09161373227834702, -0.05346088483929634, 0.07584802061319351, -0.010030749253928661, -0.0798933282494545, 0.055097345262765884, 0.07179123908281326, -0.06243592128157616, 0.13876186311244965, 0.08961252123117447, 0.08889518678188324, 0.0457143634557724, -0.002241661539301276, -0.16259117424488068, 0.029431592673063278, 0.05408640578389168, -0.0005801114020869136, 0.06688793003559113, 0.07361011207103729, 0.009169524535536766, 0.06004692614078522, -0.11718607693910599, -0.05020121857523918, 0.04235437512397766, -0.09698497503995895, -0.15449434518814087, -0.035471126437187195, 0.00028536043828353286, 0.008501144126057625, 0.0377751924097538, -0.019990913569927216, 0.05020643398165703, -0.08735095709562302, 0.07750385254621506, 0.2584896981716156, -0.2576258182525635, 0.020000271499156952, 0.060056205838918686, -0.04144195094704628, 0.04045991972088814, -0.03658195585012436, 0.02956797368824482, 0.08890560269355774, 0.027046672999858856, 0.047541916370391846, -0.11532142013311386, -0.03828473761677742, -0.015277290716767311, -0.03962590545415878, 0.013142379932105541, 0.3254058063030243, 0.06625739485025406, -0.018586471676826477, -0.010184717364609241, -0.07507702708244324, 0.037434596568346024, 0.010930763557553291, 0.043387699872255325, -0.01461172103881836, -0.027086341753602028, 0.07853902876377106, -0.0038494819309562445, -0.07637430727481842, -0.030317120254039764, -0.11660344153642654, 0.17918694019317627, 0.05456462502479553, 0.012495651841163635, -0.05226105824112892, 0.0507178008556366, -0.08664968609809875, -0.13400952517986298, -0.021491840481758118, -0.027290627360343933, -0.02244562655687332, -0.003493821481242776, -0.06132597103714943, -0.0010334925027564168, 0.11621230095624924, 0.07505137473344803, -0.16238117218017578, 0.03708553686738014, 0.01742064394056797, 0.0737495943903923, 0.022926554083824158, -0.0451202429831028, -0.11375240236520767, 0.029817353934049606, 0.12991224229335785, 0.04710274189710617, 0.08804670721292496, -0.02306073158979416, -0.078858382999897, -0.10623098164796829, 0.0869646742939949, 0.024542951956391335, -0.06951602548360825, 0.08008795231580734, -0.008666500449180603, 0.009714165702462196, -0.00021644901426043361, -0.12760454416275024, -0.011393695138394833, 0.03310149908065796, -0.025076700374484062, 0.06808990240097046, -0.002406196901574731, -0.01781788282096386, -0.0998086929321289, 0.03276912868022919, -0.08947085589170456, -0.0017633129609748721, -0.014686116948723793, -0.13084354996681213, 0.04379243776202202, -0.0384308323264122, 0.02207285352051258, -0.16235968470573425, -0.11692190915346146, 0.0038770863320678473, -0.006689894013106823, -0.04110118746757507, -0.021043356508016586, -0.010649597272276878, -0.04137226566672325, -0.0034424031618982553, -0.027185222133994102, 0.03362216800451279, -0.07690201699733734, 0.05820382386445999, 0.020469337701797485, 0.11220891773700714, -0.17032989859580994, 0.01953306794166565, -0.06791652739048004, 0.06617829948663712, -0.08170444518327713, 0.020962560549378395, 0.007337803952395916, 0.018553603440523148, -0.02126896195113659, -0.030831484124064445, -0.09114808589220047, 0.061865538358688354, 0.030641822144389153, 0.11654409766197205, -0.13365672528743744, -0.025632690638303757, 0.16577446460723877, -0.06338155269622803, -0.135273739695549, 0.19299179315567017, -0.035529352724552155, -0.03965850919485092, 0.08108481019735336, 0.10131161659955978, -0.020338764414191246, -0.02078014239668846, -0.01558306161314249, 0.027569932863116264, 0.03413376212120056, -0.05608927458524704, 0.005002449732273817, 0.009165985509753227, -0.09838941693305969, 0.03745527192950249, 0.036920372396707535, 0.036501672118902206, -0.010173502378165722, -0.009904184378683567, -0.07303299009799957, -0.028683165088295937, -0.045087460428476334, -0.02279622107744217, 0.02311266027390957, -0.1096026599407196, -0.02437184564769268, -0.11157064884901047, 0.05306416004896164, -0.06939318031072617, 0.014804905280470848, 0.007130852900445461, 0.13267505168914795, -0.05395784229040146, 0.05441569164395332, -0.1709916889667511, -0.009160680696368217, 0.009677856229245663, 0.054126136004924774, 0.03966118022799492, 0.1733521670103073, -0.008842138573527336, -0.0001020725176203996, -0.055662598460912704, 0.011407575570046902, 0.0759764239192009, -0.002962354337796569, -0.05152326822280884, -0.13386009633541107, 0.07145500183105469, -0.022823994979262352, 0.10094244033098221, -0.23715585470199585, 0.031122440472245216, -0.005130487494170666, 0.10430262982845306, 0.015929175540804863, 0.08669388294219971, -0.0015675900503993034, 0.04673285037279129, -0.11617964506149292, 0.0031974087469279766, 0.03963685408234596, -0.012508158572018147, -0.04496387392282486, 0.012545070610940456, -0.16969585418701172, 0.10410759598016739, 0.13314570486545563, -0.19655859470367432, -0.01977849006652832, -0.035020265728235245, 0.011866331100463867, 0.011249837465584278, -0.10491238534450531, -0.09159956872463226, 0.11361415684223175, -0.0035342525225132704, 0.1483946144580841, -0.13327626883983612, -0.03106085956096649, 0.01713230088353157, -0.09344294667243958, 0.010228730738162994, 0.09608194231987, 0.11048144847154617, -0.1542472392320633, 0.07596060633659363, 0.15900172293186188, -0.1670820415019989, 0.18296092748641968, 0.015204980038106441, -0.06848616153001785, 0.03282119333744049, 0.05113748088479042, -0.013647206127643585, 0.08559715747833252, -0.1086583361029625, -0.017093181610107422, 0.052077069878578186, 0.031806107610464096, 0.020615529268980026, -0.17231251299381256, 0.0025943112559616566, -0.04362226650118828, -0.046788863837718964, -0.0002744150988291949, 0.022259792312979698, 0.05601894110441208, 0.179612398147583, -0.046832140535116196, -0.10629159957170486, 0.031660594046115875, -0.03037865459918976, -0.11684370040893555, 0.21429027616977692, -0.03265342861413956, -0.2416543811559677, -0.09975466877222061, -0.053923074156045914, -0.14390473067760468, 0.04046933725476265, 0.032677046954631805, -0.0752786248922348, -0.058166638016700745, -0.10587050765752792, 0.04801848530769348, 0.03277521952986717, 0.025684116408228874, -0.050655387341976166, 0.06398683786392212, 0.06631239503622055, -0.08133814483880997, -0.057243071496486664, -0.034536585211753845, 0.03928221017122269, 0.11324037611484528, -0.04237242788076401, 0.06354785710573196, 0.099838487803936, -0.017746008932590485, 0.002520063892006874, 0.016784165054559708, 0.1377698928117752, -0.014451968483626842, 0.06146768108010292, 0.15454785525798798, -0.003070733044296503, 0.030988076701760292, 0.19246958196163177, 0.01853821985423565, -0.08514491468667984, 0.02281230129301548, 0.0727606862783432, -0.05759164318442345, -0.18596957623958588, -0.03928637132048607, -0.10781331360340118, -0.039432428777217865, 0.08943931013345718, 0.038777824491262436, 0.03296064957976341, 0.07931848615407944, -0.06912070512771606, 0.14395780861377716, 0.009714464657008648, 0.03957977890968323, 0.2391381859779358, 0.015020621009171009, 0.09866143018007278, -0.05285164341330528, -0.1012047603726387, 0.04133226349949837, 0.06859604269266129, 0.1865803748369217, -0.041349977254867554, 0.13125473260879517, 0.028744742274284363, 0.029288092628121376, 0.07461656630039215, 0.13728833198547363, 0.00927182286977768, -0.03493227809667587, -0.02410503476858139, -0.036128364503383636, -0.021687986329197884, 0.04251706227660179, -0.08573839068412781, 0.010681455954909325, -0.061559680849313736, 0.05281470715999603, 0.0551740899682045, 0.131137415766716, 0.14605239033699036, -0.2859814763069153, 0.03241302818059921, 0.0625695213675499, -0.0025521402712911367, -0.08350274711847305, 0.048697978258132935, 0.15574383735656738, -0.00107892497908324, 0.0888667032122612, 0.0227578803896904, 0.06500503420829773, -0.15531601011753082, 0.07423920929431915, -0.08527310937643051, 0.10914645344018936, -0.02372189797461033, 0.08670981973409653, -0.25874966382980347, 0.2131665050983429, 0.04665796458721161, 0.014130182564258575, -0.10785575211048126, -0.007081650663167238, 0.05038426071405411, 0.11313608288764954, 0.04488370567560196, 0.014888279139995575, -0.06266970932483673, -0.12801069021224976, -0.017668167129158974, 0.07071087509393692, 0.0268938597291708, 0.035831574350595474, 0.09314286708831787, -0.00720985047519207, 0.006084991618990898, 0.02952059730887413, -0.019767332822084427, -0.16371028125286102, -0.11608143150806427, 0.023786725476384163, 0.14993035793304443, -0.05679887905716896, -0.04789787530899048, -0.02459309808909893, -0.11677592992782593, 0.20552048087120056, -0.07033871859312057, -0.09729855507612228, -0.08694334328174591, 0.11273942142724991, -0.009208565577864647, -0.09543697535991669, -0.09276492893695831, -0.04771505668759346, 0.1051936224102974, -0.058010924607515335, -0.09200185537338257, 0.10500617325305939, -0.05568106845021248, -0.08664634823799133, -0.07139798998832703, 0.13377225399017334, -0.0464264452457428, 0.015957821160554886, 0.052625324577093124, -0.036292172968387604, 0.004902612417936325, -0.12028598040342331, 0.03128547593951225, 0.004446464125066996, 0.04257144778966904, 0.06704302132129669, -0.14481180906295776, -0.027753720059990883, -0.027229439467191696, 0.024572135880589485, 0.1676306426525116, 0.2779914140701294, -0.014178730547428131, -0.0755951851606369, 0.17744415998458862, -0.04273013398051262, -0.21093806624412537, -0.03020784631371498, -0.07480210065841675, 0.02239554561674595, -0.06379276514053345, -0.13600461184978485, 0.14106550812721252, 0.16744737327098846, 0.021056832745671272, 0.18818025290966034, -0.07055767625570297, -0.11829501390457153, 0.15215931832790375, 0.09190808981657028, 0.3067269027233124, -0.12097245454788208, -0.02853993885219097, -0.12686589360237122, -0.050140026956796646, 0.13441115617752075, -0.1569395661354065, 0.09335359185934067, -0.03604361042380333, 0.042942870408296585, 0.02007584273815155, -0.02407584711909294, 0.13464640080928802, 0.0135648837313056, 0.11973989009857178, -0.07822180539369583, 0.003527972614392638, 0.05014298856258392, -0.053194694221019745, 0.11131449043750763, -0.1656058430671692, -0.022605592384934425, -0.06992537528276443, -0.029078427702188492, -0.0494859404861927, 0.06637360900640488, -0.023502660915255547, -0.048906952142715454, -0.0530523844063282, -0.025620173662900925, -0.02374964766204357, 0.001565153943374753, 0.15293121337890625, -0.05673357844352722, 0.10047431290149689, 0.12351856380701065, 0.1657412350177765, -0.22728171944618225, 0.07549778372049332, 0.04339982196688652, -0.021812669932842255, 0.09185078740119934, -0.09620223939418793, 0.08199408650398254, 0.08220447599887848, -0.0014442646643146873, 0.1343335509300232, 0.05961444228887558, -0.04570658504962921, -0.005750601179897785, 0.07864218205213547, -0.20759017765522003, 0.010201678611338139, -0.021939946338534355, -0.03442643955349922, -0.06105216220021248, 0.016235237941145897, 0.1666848361492157, -0.004640387836843729, 0.02048570103943348, 0.006259141024202108, 0.019441967830061913, 0.005534300580620766, 0.11852747201919556, 0.08083612471818924, 0.07138516753911972, -0.09133518487215042, 0.0628134161233902, 0.0314139761030674, -0.0460246205329895, 0.021019482985138893, 0.08799371868371964, -0.12589013576507568, -0.09075721353292465, -0.09935448318719864, 0.15545792877674103, -0.0380433090031147, -0.11433029919862747, -0.15671317279338837, -0.21259315311908722, 0.030979709699749947, 0.06370222568511963, 0.06055278331041336, -0.0014987781178206205, 0.015898671001195908, -0.11465056985616684, -0.09377524256706238, 0.090199314057827, -0.02970755659043789, 0.07185933738946915, -0.1264428198337555, -0.027721285820007324, -0.023199521005153656, 0.05472993850708008, -0.03406853228807449, 0.011779583059251308, -0.13098040223121643, -0.014779789373278618, -0.2475755363702774, 0.05342628061771393, -0.09191083908081055, -0.006478320807218552, 0.015533837489783764, 0.007203012239187956, -0.06024143099784851, -0.012769668363034725, -0.06859419494867325, -0.022111693397164345, 0.0070449658669531345, 0.05312789976596832, -0.053121089935302734, -0.06919927895069122, -0.010866953991353512, -0.07971677929162979, 0.07838112115859985, -0.020385218784213066, -0.08692631125450134, 0.03496056795120239, -0.11914128810167313, -0.021415958181023598, 0.12854933738708496, 0.06750146299600601, 0.04817969352006912, -0.11559516191482544, 0.023053428158164024, 0.07575315982103348, 0.06717072427272797, 0.022950444370508194, -0.044459931552410126, -0.07643468677997589, 0.034207314252853394, -0.07744061946868896, -0.06888660043478012, -0.045123614370822906, 0.011108439415693283, 0.11984764039516449, 0.032927729189395905, 0.16070504486560822, -0.07095346599817276, -0.015101018361747265, -0.1595841944217682, 0.03796066716313362, 0.01608196645975113, -0.13250601291656494, 0.03017246164381504, -0.02623756229877472, 0.0690075159072876, -0.012302366085350513, 0.23745091259479523, -0.02097632922232151, -0.024434275925159454, 0.025467243045568466, 0.021420607343316078, 0.0018594436114653945, 0.018035592511296272, 0.15030381083488464, 0.015390514396131039, -0.061392441391944885, -0.0847509577870369, 0.007937673479318619, 0.1717982441186905, 0.15343095362186432, 0.2311018705368042, 0.0825878158211708, 0.019423460587859154, 0.10082750767469406, 0.05277590826153755, -0.08419443666934967, 0.10583683103322983, 0.10626349598169327, -0.00236330134794116, 0.07927757501602173, 0.03275909647345543, -0.005786839872598648, 0.2353716641664505, -0.05662713199853897, -0.04177609458565712, -0.04966024309396744, -0.07852394133806229, -0.14763936400413513, -0.18077808618545532, -0.1143682673573494, -0.0623214952647686, -0.03177618980407715, -0.12323243916034698, -0.06908102333545685, 0.11599893867969513, 0.05165030062198639, -0.016168110072612762, 0.03245440497994423, 0.062401771545410156, -0.08065873384475708, 0.07078361511230469, 0.01260558795183897, 0.013449834659695625, -0.0007582195103168488, -0.07404522597789764, 0.038541246205568314, -0.05680086091160774, -0.01292529795318842, 0.027682049199938774, 0.04857155680656433, 0.027943408116698265, -0.09598246961832047, -0.06969083845615387, -0.019639888778328896, 0.023519139736890793, 0.02216988615691662, 0.12693212926387787, 0.06600118428468704, -0.06071804091334343, 0.03311210870742798, 0.058047398924827576, -0.023898636922240257, -0.027118835598230362, -0.09571376442909241, 0.12123259156942368, -0.03539891168475151, 0.06279777735471725, -0.0007737912819720805, -0.08735650032758713, 0.031279437243938446, 0.19867637753486633, 0.19663143157958984, -0.06298746913671494, 0.030656056478619576, -0.08467244356870651, -0.00257022213190794, -0.0033736054319888353, 0.08328059315681458, 0.08711683005094528, 0.11538172513246536, -0.009049873799085617, -0.068326435983181, -0.021897634491324425, 0.02805158868432045, -0.12552061676979065, 0.04485237970948219, 0.0570005439221859, 0.00023666345805395395, -0.05960026755928993, 0.009237471967935562, -0.06579340994358063, -0.02306373231112957, 0.00006015569670125842, -0.04003873094916344, -0.0818549245595932, -0.013278855942189693, -0.06250254064798355, -0.027402959764003754, 0.057945676147937775, -0.11665424704551697, -0.04055061936378479, 0.0626051053404808, -0.03459746390581131, -0.0967266857624054, -0.028257112950086594, 0.029242651537060738, 0.05213049054145813, 0.02761266939342022, 0.009821149520576, 0.1276916116476059, 0.10875432938337326, -0.008793496526777744, -0.07161937654018402, 0.13742657005786896, 0.01274383906275034, 0.04583129286766052, 0.047315564006567, 0.07619228214025497, 0.032868482172489166, 0.003549129469320178, 0.12409251928329468, -0.041561488062143326, -0.001082056318409741, -0.042340539395809174, -0.1015012115240097, -0.08586864173412323, 0.01180165633559227, -0.060944266617298126, 0.08950544148683548, 0.1042010486125946, -0.03248070180416107, -0.02462414652109146, -0.021301599219441414, -0.01198319811373949, 0.04666238650679588, -0.10871363431215286, -0.018622754141688347, -0.17386394739151, 0.040515605360269547, 0.08482182025909424, -0.014823570847511292, -0.24811400473117828, -0.010013008490204811, -0.06254889816045761, -0.030145976692438126, -0.06433442234992981, 0.037734851241111755, 0.11421982198953629, 0.07526279240846634, -0.07277046889066696, -0.16192665696144104, -0.028589263558387756, 0.10587073117494583, -0.07714030146598816, -0.1375417709350586 ]
null
null
null
# 基于SDXL模型LoRA微调实现《少前2:追放》文生图 ![example](./assets/example.png) ## 一、Model Library 1. 微调数据集:[基于SDXL模型的《少女前线2:追放》LoRA微调数据集](https://www.kaggle.com/datasets/yukikonata/sdxl2lora) 2. 预训练模型:[stable_diffusion_xl](https://huggingface.co/docs/diffusers/api/pipelines/stable_diffusion/stable_diffusion_xl) 3. 底模:[animagine-xl-3.0](https://huggingface.co/cagliostrolab/animagine-xl-3.0) 4. SDXL LoRA微调训练器:[kohya_ss](https://github.com/bmaltais/kohya_ss) 5. 数据集画质增强:[waifu2x](https://github.com/nagadomi/waifu2x) ## 二、Prompt Dict 1. 少前2追放角色 * 佩里缇亚: PKPSP * 塞布丽娜: SPAS12 * 托洛洛: AKAlfa * 桑朵莱希: G36 * 琼玖: QBZ191 * 维普雷: Vepr12 * 莫辛纳甘: MosinNagant * 黛烟: QBZ95 * 克罗丽科: Kroliko * 夏克里: XCRL * 奇塔: MP7 * 寇尔芙: TaurusCurve * 科谢尼娅: APS * 纳甘: Nagant1895 * 纳美西丝: OM50 * 莉塔拉: GalilARM * 闪电: OTs14 2. Pixiv画师风格 * おにねこ(鬼猫): Onineko26 * 麻生: AsouAsabu * mignon: Mignon * migolu: Migolu ## 三、使用方式 1. 安装部分环境(默认已安装pytorch等必要环境) ~~~sh pip install diffusers --upgrade pip install transformers accelerate safetensors ~~~ 2. 使用Hugging Face下载并使用底模(animagine-xl-3.0)和LoRA模型 ~~~python import torch import datetime from PIL import Image import matplotlib.pyplot as plt from diffusers import ( StableDiffusionXLPipeline, EulerAncestralDiscreteScheduler, AutoencoderKL ) # LoRA Hugging Face ID lora_id = "TfiyuenLau/GirlsFrontline2_SDXL_LoRA" # Load VAE component vae = AutoencoderKL.from_pretrained( "madebyollin/sdxl-vae-fp16-fix", torch_dtype=torch.float16 ) # Configure the pipeline pipe = StableDiffusionXLPipeline.from_pretrained( "cagliostrolab/animagine-xl-3.0", vae=vae, torch_dtype=torch.float16, use_safetensors=True, ) pipe.load_lora_weights(lora_id) pipe.scheduler = EulerAncestralDiscreteScheduler.from_config(pipe.scheduler.config) pipe.to('cuda') ~~~ 3. 生成图像 ~~~python # Define Prompt output = "./output.png" prompt = "1girl, OTs14, gloves, looking at viewer, smile, food, holding, solo, closed mouth, sitting, yellow eyes, black gloves, masterpiece, best quality" negative_prompt = "nsfw, lowres, bad anatomy, bad hands, text, error, missing fingers, extra digit, fewer digits, cropped, worst quality, low quality, normal quality, jpeg artifacts, signature, watermark, username, blurry, artist name" # Generate Image image = pipe( prompt, negative_prompt=negative_prompt, width=1024, height=1024, guidance_scale=7, num_inference_steps=28 ).images[0] # Save & Show image.save(output) image = Image.open(output) plt.axis('off') plt.imshow(image) image.close() ~~~
{"language": "en", "license": "apache-2.0", "tags": ["text-to-image", "diffusion models", "LoRA fine-tuning", "animagine-xl-3.0", "stable_diffusion_xl", "kohya_ss", "waifu2x"], "thumbnail": "./assets/example.png", "model": {"name": "GirlsFrontline2-SDXL-LoRA", "description": "A model based on SDXL with LoRA fine-tuning for \u300aGirl's Frontline 2\u300b Text-To-Image generation.", "pipeline_tag": "image-generation", "repo": "https://huggingface.co/TfiyuenLau/GirlsFrontline2_SDXL_LoRA", "library": "huggingface", "framework": "pytorch", "version": "1.0.0", "pretrained_model": "stable_diffusion_xl", "base_model": "animagine-xl-3.0", "fine_tuner": "kohya_ss", "data_augmentation": "waifu2x", "task": "text-to-image"}}
text-to-image
TfiyuenLau/GirlsFrontline2_SDXL_LoRA
[ "tensorboard", "text-to-image", "diffusion models", "LoRA fine-tuning", "animagine-xl-3.0", "stable_diffusion_xl", "kohya_ss", "waifu2x", "en", "license:apache-2.0", "region:us" ]
2024-02-10T12:30:23+00:00
[]
[ "en" ]
TAGS #tensorboard #text-to-image #diffusion models #LoRA fine-tuning #animagine-xl-3.0 #stable_diffusion_xl #kohya_ss #waifu2x #en #license-apache-2.0 #region-us
# 基于SDXL模型LoRA微调实现《少前2:追放》文生图 !example ## 一、Model Library 1. 微调数据集:基于SDXL模型的《少女前线2:追放》LoRA微调数据集 2. 预训练模型:stable_diffusion_xl 3. 底模:animagine-xl-3.0 4. SDXL LoRA微调训练器:kohya_ss 5. 数据集画质增强:waifu2x ## 二、Prompt Dict 1. 少前2追放角色 * 佩里缇亚: PKPSP * 塞布丽娜: SPAS12 * 托洛洛: AKAlfa * 桑朵莱希: G36 * 琼玖: QBZ191 * 维普雷: Vepr12 * 莫辛纳甘: MosinNagant * 黛烟: QBZ95 * 克罗丽科: Kroliko * 夏克里: XCRL * 奇塔: MP7 * 寇尔芙: TaurusCurve * 科谢尼娅: APS * 纳甘: Nagant1895 * 纳美西丝: OM50 * 莉塔拉: GalilARM * 闪电: OTs14 2. Pixiv画师风格 * おにねこ(鬼猫): Onineko26 * 麻生: AsouAsabu * mignon: Mignon * migolu: Migolu ## 三、使用方式 1. 安装部分环境(默认已安装pytorch等必要环境) ~~~sh pip install diffusers --upgrade pip install transformers accelerate safetensors ~~~ 2. 使用Hugging Face下载并使用底模(animagine-xl-3.0)和LoRA模型 ~~~python import torch import datetime from PIL import Image import URL as plt from diffusers import ( StableDiffusionXLPipeline, EulerAncestralDiscreteScheduler, AutoencoderKL ) # LoRA Hugging Face ID lora_id = "TfiyuenLau/GirlsFrontline2_SDXL_LoRA" # Load VAE component vae = AutoencoderKL.from_pretrained( "madebyollin/sdxl-vae-fp16-fix", torch_dtype=torch.float16 ) # Configure the pipeline pipe = StableDiffusionXLPipeline.from_pretrained( "cagliostrolab/animagine-xl-3.0", vae=vae, torch_dtype=torch.float16, use_safetensors=True, ) pipe.load_lora_weights(lora_id) pipe.scheduler = EulerAncestralDiscreteScheduler.from_config(URL) URL('cuda') ~~~ 3. 生成图像 ~~~python # Define Prompt output = "./URL" prompt = "1girl, OTs14, gloves, looking at viewer, smile, food, holding, solo, closed mouth, sitting, yellow eyes, black gloves, masterpiece, best quality" negative_prompt = "nsfw, lowres, bad anatomy, bad hands, text, error, missing fingers, extra digit, fewer digits, cropped, worst quality, low quality, normal quality, jpeg artifacts, signature, watermark, username, blurry, artist name" # Generate Image image = pipe( prompt, negative_prompt=negative_prompt, width=1024, height=1024, guidance_scale=7, num_inference_steps=28 ).images[0] # Save & Show URL(output) image = URL(output) URL('off') URL(image) URL() ~~~
[ "# 基于SDXL模型LoRA微调实现《少前2:追放》文生图\n\n!example", "## 一、Model Library\n1. 微调数据集:基于SDXL模型的《少女前线2:追放》LoRA微调数据集\n2. 预训练模型:stable_diffusion_xl\n3. 底模:animagine-xl-3.0\n4. SDXL LoRA微调训练器:kohya_ss\n5. 数据集画质增强:waifu2x", "## 二、Prompt Dict\n1. 少前2追放角色\n* 佩里缇亚: PKPSP\n* 塞布丽娜: SPAS12\n* 托洛洛: AKAlfa\n* 桑朵莱希: G36\n* 琼玖: QBZ191\n* 维普雷: Vepr12\n* 莫辛纳甘: MosinNagant\n* 黛烟: QBZ95\n* 克罗丽科: Kroliko\n* 夏克里: XCRL\n* 奇塔: MP7\n* 寇尔芙: TaurusCurve\n* 科谢尼娅: APS\n* 纳甘: Nagant1895\n* 纳美西丝: OM50\n* 莉塔拉: GalilARM\n* 闪电: OTs14\n\n2. Pixiv画师风格\n* おにねこ(鬼猫): Onineko26\n* 麻生: AsouAsabu\n* mignon: Mignon\n* migolu: Migolu", "## 三、使用方式\n1. 安装部分环境(默认已安装pytorch等必要环境)\n~~~sh\npip install diffusers --upgrade\npip install transformers accelerate safetensors\n~~~\n\n2. 使用Hugging Face下载并使用底模(animagine-xl-3.0)和LoRA模型\n~~~python\nimport torch\nimport datetime\nfrom PIL import Image\nimport URL as plt\nfrom diffusers import (\n StableDiffusionXLPipeline, \n EulerAncestralDiscreteScheduler,\n AutoencoderKL\n)", "# LoRA Hugging Face ID\nlora_id = \"TfiyuenLau/GirlsFrontline2_SDXL_LoRA\"", "# Load VAE component\nvae = AutoencoderKL.from_pretrained(\n \"madebyollin/sdxl-vae-fp16-fix\", \n torch_dtype=torch.float16\n)", "# Configure the pipeline\npipe = StableDiffusionXLPipeline.from_pretrained(\n \"cagliostrolab/animagine-xl-3.0\", \n vae=vae,\n torch_dtype=torch.float16, \n use_safetensors=True, \n)\npipe.load_lora_weights(lora_id)\npipe.scheduler = EulerAncestralDiscreteScheduler.from_config(URL)\nURL('cuda')\n~~~\n\n3. 生成图像\n~~~python", "# Define Prompt\noutput = \"./URL\"\nprompt = \"1girl, OTs14, gloves, looking at viewer, smile, food, holding, solo, closed mouth, sitting, yellow eyes, black gloves, masterpiece, best quality\"\nnegative_prompt = \"nsfw, lowres, bad anatomy, bad hands, text, error, missing fingers, extra digit, fewer digits, cropped, worst quality, low quality, normal quality, jpeg artifacts, signature, watermark, username, blurry, artist name\"", "# Generate Image\nimage = pipe(\n prompt, \n negative_prompt=negative_prompt, \n width=1024,\n height=1024,\n guidance_scale=7,\n num_inference_steps=28\n).images[0]", "# Save & Show\nURL(output)\nimage = URL(output)\nURL('off')\nURL(image)\nURL()\n~~~" ]
[ "TAGS\n#tensorboard #text-to-image #diffusion models #LoRA fine-tuning #animagine-xl-3.0 #stable_diffusion_xl #kohya_ss #waifu2x #en #license-apache-2.0 #region-us \n", "# 基于SDXL模型LoRA微调实现《少前2:追放》文生图\n\n!example", "## 一、Model Library\n1. 微调数据集:基于SDXL模型的《少女前线2:追放》LoRA微调数据集\n2. 预训练模型:stable_diffusion_xl\n3. 底模:animagine-xl-3.0\n4. SDXL LoRA微调训练器:kohya_ss\n5. 数据集画质增强:waifu2x", "## 二、Prompt Dict\n1. 少前2追放角色\n* 佩里缇亚: PKPSP\n* 塞布丽娜: SPAS12\n* 托洛洛: AKAlfa\n* 桑朵莱希: G36\n* 琼玖: QBZ191\n* 维普雷: Vepr12\n* 莫辛纳甘: MosinNagant\n* 黛烟: QBZ95\n* 克罗丽科: Kroliko\n* 夏克里: XCRL\n* 奇塔: MP7\n* 寇尔芙: TaurusCurve\n* 科谢尼娅: APS\n* 纳甘: Nagant1895\n* 纳美西丝: OM50\n* 莉塔拉: GalilARM\n* 闪电: OTs14\n\n2. Pixiv画师风格\n* おにねこ(鬼猫): Onineko26\n* 麻生: AsouAsabu\n* mignon: Mignon\n* migolu: Migolu", "## 三、使用方式\n1. 安装部分环境(默认已安装pytorch等必要环境)\n~~~sh\npip install diffusers --upgrade\npip install transformers accelerate safetensors\n~~~\n\n2. 使用Hugging Face下载并使用底模(animagine-xl-3.0)和LoRA模型\n~~~python\nimport torch\nimport datetime\nfrom PIL import Image\nimport URL as plt\nfrom diffusers import (\n StableDiffusionXLPipeline, \n EulerAncestralDiscreteScheduler,\n AutoencoderKL\n)", "# LoRA Hugging Face ID\nlora_id = \"TfiyuenLau/GirlsFrontline2_SDXL_LoRA\"", "# Load VAE component\nvae = AutoencoderKL.from_pretrained(\n \"madebyollin/sdxl-vae-fp16-fix\", \n torch_dtype=torch.float16\n)", "# Configure the pipeline\npipe = StableDiffusionXLPipeline.from_pretrained(\n \"cagliostrolab/animagine-xl-3.0\", \n vae=vae,\n torch_dtype=torch.float16, \n use_safetensors=True, \n)\npipe.load_lora_weights(lora_id)\npipe.scheduler = EulerAncestralDiscreteScheduler.from_config(URL)\nURL('cuda')\n~~~\n\n3. 生成图像\n~~~python", "# Define Prompt\noutput = \"./URL\"\nprompt = \"1girl, OTs14, gloves, looking at viewer, smile, food, holding, solo, closed mouth, sitting, yellow eyes, black gloves, masterpiece, best quality\"\nnegative_prompt = \"nsfw, lowres, bad anatomy, bad hands, text, error, missing fingers, extra digit, fewer digits, cropped, worst quality, low quality, normal quality, jpeg artifacts, signature, watermark, username, blurry, artist name\"", "# Generate Image\nimage = pipe(\n prompt, \n negative_prompt=negative_prompt, \n width=1024,\n height=1024,\n guidance_scale=7,\n num_inference_steps=28\n).images[0]", "# Save & Show\nURL(output)\nimage = URL(output)\nURL('off')\nURL(image)\nURL()\n~~~" ]
[ 65, 26, 84, 213, 121, 33, 52, 125, 127, 58, 31 ]
[ "passage: TAGS\n#tensorboard #text-to-image #diffusion models #LoRA fine-tuning #animagine-xl-3.0 #stable_diffusion_xl #kohya_ss #waifu2x #en #license-apache-2.0 #region-us \n# 基于SDXL模型LoRA微调实现《少前2:追放》文生图\n\n!example## 一、Model Library\n1. 微调数据集:基于SDXL模型的《少女前线2:追放》LoRA微调数据集\n2. 预训练模型:stable_diffusion_xl\n3. 底模:animagine-xl-3.0\n4. SDXL LoRA微调训练器:kohya_ss\n5. 数据集画质增强:waifu2x## 二、Prompt Dict\n1. 少前2追放角色\n* 佩里缇亚: PKPSP\n* 塞布丽娜: SPAS12\n* 托洛洛: AKAlfa\n* 桑朵莱希: G36\n* 琼玖: QBZ191\n* 维普雷: Vepr12\n* 莫辛纳甘: MosinNagant\n* 黛烟: QBZ95\n* 克罗丽科: Kroliko\n* 夏克里: XCRL\n* 奇塔: MP7\n* 寇尔芙: TaurusCurve\n* 科谢尼娅: APS\n* 纳甘: Nagant1895\n* 纳美西丝: OM50\n* 莉塔拉: GalilARM\n* 闪电: OTs14\n\n2. Pixiv画师风格\n* おにねこ(鬼猫): Onineko26\n* 麻生: AsouAsabu\n* mignon: Mignon\n* migolu: Migolu" ]
[ -0.01026773452758789, 0.12035059928894043, -0.013864881359040737, 0.0660625770688057, 0.0897306427359581, 0.01106198225170374, 0.0713854655623436, 0.10473394393920898, 0.05220695585012436, 0.09146631509065628, 0.06294351071119308, 0.1170414388179779, 0.09398120641708374, 0.13552001118659973, -0.009199519641697407, -0.27518680691719055, 0.036263443529605865, -0.011130673810839653, -0.06117238849401474, 0.05533680319786072, 0.06250035762786865, -0.04758439213037491, 0.08277592808008194, -0.036427225917577744, -0.0014468879671767354, -0.053370799869298935, -0.03497495874762535, -0.05014447495341301, 0.030910922214388847, 0.07076995819807053, 0.01373730693012476, 0.04874415323138237, 0.012273464351892471, -0.15134471654891968, 0.0344514362514019, 0.0012418415863066912, -0.06513547897338867, 0.0024643559008836746, 0.06685794144868851, -0.03887559846043587, 0.18834511935710907, -0.08496198058128357, 0.0038483920507133007, -0.0023657565470784903, -0.10909979790449142, -0.09076891839504242, -0.027096012607216835, 0.09130001813173294, 0.10371432453393936, 0.019782908260822296, 0.003797368612140417, 0.06443315744400024, -0.08497168123722076, 0.07974950969219208, 0.19487425684928894, -0.20572809875011444, -0.0964435264468193, 0.04550573602318764, 0.09867745637893677, 0.04993200674653053, -0.08581695705652237, 0.03581388667225838, -0.006140802055597305, -0.013098110444843769, -0.017367737367749214, -0.06859049946069717, 0.14888639748096466, -0.041982922703027725, -0.09093288332223892, 0.00396769167855382, 0.1239418089389801, 0.04434146359562874, -0.005972878076136112, -0.03890962898731232, -0.0035767925437539816, -0.010187598876655102, -0.0865035355091095, -0.011721526272594929, 0.01563429646193981, -0.043193526566028595, 0.07152368873357773, 0.0046769133768975735, -0.03865864872932434, -0.016868017613887787, 0.016058070585131645, 0.10181083530187607, 0.026752101257443428, -0.008767761290073395, 0.012563737109303474, -0.024672582745552063, -0.054224226623773575, -0.1309456080198288, -0.0054114204831421375, -0.046520672738552094, -0.015238821506500244, 0.04392688721418381, 0.04428192228078842, 0.0419464148581028, 0.13224102556705475, 0.148273766040802, -0.014084714464843273, 0.10259407758712769, 0.030926669016480446, 0.024978501722216606, -0.000675623188726604, 0.015220416709780693, -0.06069076061248779, -0.10300860553979874, -0.05046696215867996, 0.07698048651218414, -0.004900132771581411, -0.07485492527484894, -0.0591670423746109, 0.008926153182983398, -0.043167609721422195, 0.023802539333701134, 0.07350579649209976, 0.032944079488515854, -0.027449287474155426, -0.04031113162636757, 0.09545642137527466, -0.04851172864437103, 0.049032747745513916, 0.06151566654443741, -0.02397693134844303, 0.044148750603199005, -0.04976288601756096, 0.0447448268532753, -0.011084282770752907, 0.06828484684228897, -0.03643645718693733, 0.008052356541156769, -0.018951769918203354, -0.025194572284817696, 0.03434751182794571, -0.04215211421251297, 0.022077465429902077, -0.12554709613323212, 0.011037093587219715, -0.068915456533432, 0.04437598958611488, -0.07395344227552414, -0.0431380569934845, 0.00367920845746994, -0.08988561481237411, 0.05783711373806, 0.03245740011334419, 0.04767415672540665, -0.039427757263183594, 0.04621677100658417, -0.017603803426027298, 0.11054757237434387, -0.0062012821435928345, 0.02928253635764122, 0.010801327414810658, 0.05862528830766678, -0.14225316047668457, -0.0076343645341694355, -0.09775004535913467, 0.024318790063261986, -0.07336172461509705, -0.04316047206521034, -0.028650274500250816, 0.005149006377905607, 0.03184167295694351, 0.08459743857383728, -0.15362651646137238, -0.03012046217918396, 0.11288570612668991, -0.07694632560014725, -0.07786426693201065, 0.08955377340316772, 0.04503515362739563, -0.014194093644618988, 0.015334864147007465, 0.245462104678154, 0.07455413043498993, -0.04915091395378113, -0.05175646021962166, -0.04642634093761444, 0.007115129381418228, 0.07872059941291809, 0.11231625825166702, 0.03817886859178543, 0.07189729064702988, 0.0298441331833601, -0.11034727096557617, -0.007019815035164356, -0.03530114144086838, -0.05729180946946144, 0.025755006819963455, -0.0630323737859726, 0.012400654144585133, 0.05639045685529709, -0.007192676886916161, -0.009956675581634045, -0.03556206077337265, -0.0775797888636589, 0.07893671095371246, 0.0039053475484251976, 0.009680015034973621, -0.07914473116397858, 0.14696860313415527, 0.11896708607673645, 0.02456807903945446, -0.07626847922801971, 0.02666061744093895, 0.003318818286061287, 0.007486917078495026, 0.009112272411584854, -0.05267692729830742, 0.057091616094112396, 0.05097620561718941, 0.004283032845705748, -0.01920202746987343, 0.012775692157447338, 0.034135814756155014, -0.008975094184279442, -0.17399919033050537, 0.025154858827590942, -0.03475826233625412, 0.0692405253648758, -0.13211564719676971, 0.00044718809658661485, -0.001763456966727972, 0.11626292020082474, 0.010793742723762989, -0.019637716934084892, -0.0046997093595564365, 0.021166354417800903, -0.01017281599342823, 0.017010577023029327, 0.05951464921236038, -0.014463347382843494, -0.02415868267416954, 0.07313857227563858, -0.10220536589622498, 0.15659867227077484, 0.08242480456829071, 0.02137444168329239, -0.04286674037575722, -0.025342442095279694, 0.0041386960074305534, -0.003680592169985175, 0.09681256115436554, 0.010413112118840218, 0.054305288940668106, 0.010442827828228474, 0.06006820872426033, -0.06283831596374512, 0.04600122943520546, 0.054654866456985474, -0.0549173578619957, -0.032269008457660675, 0.16783367097377777, 0.13177995383739471, -0.08754777908325195, 0.05196312814950943, 0.12385556846857071, 0.0393911711871624, 0.20536258816719055, -0.009648391045629978, -0.03333967551589012, -0.11476694792509079, 0.022579729557037354, 0.04770086333155632, 0.08178563416004181, -0.08785707503557205, 0.006172165274620056, 0.020138079300522804, -0.031857848167419434, -0.01932937279343605, -0.08893939852714539, -0.0676814541220665, -0.028218235820531845, -0.01236417330801487, 0.002216621767729521, 0.06297571212053299, -0.02798066847026348, 0.1105726808309555, 0.021933339536190033, -0.09910222142934799, -0.08521951735019684, -0.02262345515191555, -0.02462955377995968, 0.1191025823354721, -0.08388116210699081, -0.1616671085357666, 0.03425158932805061, -0.06486546248197556, -0.04250463470816612, -0.018591126427054405, 0.02863125316798687, -0.05861999839544296, -0.016506711021065712, -0.036026481539011, -0.024032356217503548, -0.02410184219479561, -0.02078828401863575, 0.018152177333831787, 0.03205748647451401, -0.05771586671471596, -0.040905289351940155, -0.02263752371072769, -0.0244402177631855, -0.016340747475624084, 0.06911866366863251, -0.04133959859609604, 0.14985033869743347, 0.10607331246137619, 0.06325509399175644, 0.01903674751520157, 0.02128155529499054, 0.1777838170528412, -0.11184906959533691, 0.07432739436626434, 0.061704427003860474, 0.008285311050713062, 0.09108331054449081, 0.1739267259836197, 0.04729260131716728, -0.04294746369123459, -0.018441088497638702, 0.0031819785945117474, -0.05139630287885666, -0.10046971589326859, -0.03397805616259575, -0.03249211236834526, 0.11351663619279861, 0.006079578772187233, 0.05566805228590965, 0.02582707069814205, 0.03855254873633385, -0.006631522439420223, -0.04672468826174736, 0.057335760444402695, 0.03487008437514305, 0.02723338082432747, -0.012403919361531734, 0.06446058303117752, -0.03104802593588829, -0.014404001645743847, 0.0424346923828125, 0.021142274141311646, 0.11540006101131439, 0.04465120658278465, 0.05507444590330124, 0.0688374936580658, 0.06729521602392197, 0.015946468338370323, -0.04060951992869377, 0.04289988800883293, -0.04444418475031853, -0.024430349469184875, -0.04937160387635231, 0.028141280636191368, 0.06353665143251419, 0.10782673954963684, -0.0661497488617897, -0.06442403048276901, 0.0796029344201088, 0.06056675687432289, 0.06263981759548187, 0.004891045391559601, -0.08845819532871246, 0.008073925040662289, 0.02507849782705307, 0.012930316850543022, -0.028518181294202805, 0.004331143107265234, 0.009791775606572628, -0.09179689735174179, 0.05618255212903023, 0.02347833849489689, 0.07146745920181274, -0.0017404325772076845, 0.027022158727049828, -0.026260068640112877, -0.027400871738791466, 0.0010928417323157191, 0.021723583340644836, -0.1906731277704239, 0.1840246319770813, -0.0007847234373912215, 0.01743905618786812, -0.018098000437021255, -0.039411626756191254, 0.06445428729057312, 0.04499271884560585, 0.1437394917011261, 0.015041142702102661, -0.15400943160057068, -0.16462811827659607, -0.07655083388090134, 0.00553209288045764, 0.1121324971318245, -0.06508322805166245, 0.06442290544509888, 0.002490012440830469, -0.03180564194917679, -0.04622127488255501, 0.03634955734014511, -0.11718647181987762, -0.061335064470767975, 0.09387073665857315, -0.01835477724671364, -0.04519132897257805, -0.026292864233255386, -0.05399659276008606, -0.10395117849111557, 0.038681138306856155, -0.08003116399049759, -0.01808604970574379, -0.048972904682159424, -0.03732338920235634, 0.07390031218528748, -0.10246279090642929, 0.028243741020560265, -0.035745423287153244, 0.018869590014219284, -0.028436459600925446, -0.03840174525976181, 0.028759794309735298, -0.06726042926311493, -0.1363542079925537, -0.02934393845498562, 0.08961531519889832, 0.006744219455868006, 0.07513663172721863, 0.028919043019413948, 0.04026171192526817, 0.007045913953334093, -0.1001012772321701, 0.0833616778254509, 0.009302226826548576, -0.02858603000640869, 0.006536555010825396, -0.036079611629247665, -0.07760549336671829, -0.09305591881275177, -0.05001355707645416, 0.1000397577881813, 0.27602308988571167, -0.05473478510975838, 0.11594125628471375, 0.039905186742544174, -0.04453833028674126, -0.18590539693832397, -0.11316104233264923, 0.06355175375938416, -0.013535656034946442, 0.05557779222726822, -0.18100833892822266, 0.00838328804820776, 0.020184239372611046, -0.011316047981381416, 0.11069757491350174, -0.2837097644805908, -0.07128383219242096, 0.04103318229317665, 0.06086806580424309, 0.06018304452300072, -0.2214243859052658, -0.0714942216873169, -0.004403931088745594, -0.11787013709545135, 0.09485775232315063, 0.0185893252491951, 0.07660982012748718, -0.04681830108165741, -0.06669045239686966, 0.026264678686857224, -0.0056323944590985775, 0.18027342855930328, -0.002350982278585434, 0.002168169477954507, -0.11128893494606018, -0.15399552881717682, -0.021592913195490837, -0.03095833957195282, 0.004586121533066034, -0.0465373694896698, -0.050506699830293655, -0.15141823887825012, 0.01722741685807705, -0.10306235402822495, 0.012018414214253426, -0.062065161764621735, 0.009622197598218918, -0.04630396515130997, 0.09887776523828506, 0.04890188202261925, 0.034087780863046646, 0.03842870891094208, -0.041273970156908035, 0.03323454409837723, 0.09726306796073914, 0.029500525444746017, 0.033293113112449646, -0.08023619651794434, -0.04637310281395912, -0.016910729929804802, 0.013696350157260895, -0.09629091620445251, -0.010524329729378223, 0.1099628284573555, 0.030294738709926605, 0.06147110462188721, 0.02342129498720169, -0.0609230138361454, 0.003264992032200098, 0.1303510069847107, -0.08474859595298767, -0.11545252799987793, -0.03796955570578575, -0.018100818619132042, 0.03399846702814102, -0.05817839503288269, 0.07877730578184128, 0.0035419827327132225, -0.0058488743379712105, -0.0052450732327997684, 0.06725303828716278, 0.031004372984170914, 0.06765077263116837, 0.07454942911863327, 0.02513984777033329, -0.06491196155548096, 0.012251955457031727, 0.05522379279136658, -0.06138177216053009, 0.040376145392656326, 0.1487879902124405, -0.041249532252550125, -0.07261449098587036, 0.00856113899499178, 0.07596646994352341, -0.0006472899694927037, -0.023197295144200325, -0.0579473078250885, -0.07154539227485657, 0.033647652715444565, 0.09058351069688797, 0.015769831836223602, -0.038820136338472366, 0.05992192402482033, -0.007173479069024324, -0.006813377141952515, 0.0922149121761322, 0.041130587458610535, -0.006011488381773233, -0.07710221409797668, 0.018513968214392662, 0.01759658381342888, 0.056197114288806915, -0.008224052377045155, -0.026160014793276787, -0.14740055799484253, -0.00783529132604599, -0.044173505157232285, 0.031562261283397675, -0.12785199284553528, -0.02696765772998333, -0.021295059472322464, -0.02147722989320755, -0.06802765280008316, -0.0141407186165452, -0.05661972239613533, -0.03167913109064102, -0.025414161384105682, 0.09425871074199677, -0.06582633405923843, -0.06288502365350723, 0.05391382426023483, -0.046457357704639435, 0.023209767416119576, 0.030517999082803726, 0.020425133407115936, 0.01579592004418373, -0.10224734246730804, 0.018023665994405746, -0.03069721721112728, 0.015401553362607956, 0.00498688081279397, -0.09925468266010284, -0.002111000707373023, -0.0679822489619255, 0.0009322363184764981, 0.02916785329580307, 0.06575272977352142, -0.07319032400846481, 0.02147197723388672, -0.049267929047346115, -0.12338661402463913, -0.020742207765579224, 0.03828861191868782, 0.07842468470335007, 0.018514087423682213, 0.05732303857803345, -0.0150609714910388, 0.09470229595899582, -0.10909031331539154, 0.006235018372535706, -0.006770418491214514, -0.059224892407655716, 0.04930999502539635, -0.0353396013379097, 0.04562866687774658, -0.027295105159282684, 0.016176754608750343, 0.02585444040596485, -0.023665696382522583, -0.00983034260571003, -0.04535974562168121, -0.012147966772317886, 0.04877258837223053, 0.11195680499076843, 0.059988390654325485, -0.025220653042197227, -0.009825499728322029, 0.03538823872804642, 0.006909750401973724, 0.06411860138177872, 0.06860291212797165, 0.11277975887060165, 0.11894752830266953, 0.04146905243396759, 0.0876833125948906, -0.06653624027967453, -0.06431613117456436, 0.06661882996559143, -0.07121389359235764, 0.08484792709350586, -0.0389241985976696, -0.026305263862013817, 0.13298086822032928, -0.1454625427722931, 0.04689624160528183, -0.07213301956653595, -0.057227425277233124, -0.07068377733230591, -0.14002875983715057, -0.08275879919528961, -0.03270687907934189, 0.02505205199122429, -0.07937385141849518, 0.04570098966360092, 0.036601997911930084, 0.006823782809078693, 0.015710776671767235, 0.061871930956840515, 0.006291630212217569, -0.021404171362519264, 0.11030883342027664, -0.002832795260474086, -0.030939918011426926, 0.018663926050066948, 0.013673270121216774, -0.023303937166929245, -0.0629640743136406, -0.027514148503541946, 0.03266815096139908, -0.0748797059059143, 0.011939623393118382, -0.0427803210914135, -0.08246719092130661, 0.023464027792215347, 0.003115671453997493, 0.032434508204460144, 0.1485201120376587, 0.037704430520534515, -0.035158637911081314, 0.005019602831453085, 0.11782769858837128, 0.03229104354977608, 0.015969449654221535, -0.09158270806074142, 0.06027849018573761, -0.006352155469357967, -0.013671668246388435, -0.02718435600399971, -0.03229624032974243, -0.011524294503033161, 0.1686912626028061, 0.13898786902427673, -0.07787925004959106, 0.02708865888416767, 0.03520060330629349, 0.038951680064201355, 0.03044748120009899, 0.08903399854898453, 0.0802958682179451, 0.20119759440422058, -0.05189912021160126, -0.03433702141046524, -0.0909927487373352, -0.02684093825519085, -0.08102769404649734, 0.018060732632875443, 0.09587036818265915, -0.0031129447743296623, -0.03980820253491402, 0.11453621089458466, -0.093614362180233, -0.01185148861259222, -0.008770083077251911, -0.1874147653579712, -0.10498005896806717, -0.02496841922402382, 0.0038937360513955355, 0.08366529643535614, 0.07670841366052628, -0.015033034607768059, -0.05255461856722832, 0.00546928821131587, 0.03872887045145035, -0.07269419729709625, 0.02697594091296196, 0.045660726726055145, -0.07077524065971375, 0.15591736137866974, -0.028886841610074043, 0.06420158594846725, 0.08659029752016068, 0.013578442856669426, -0.05377679318189621, 0.04437108710408211, 0.07615645229816437, -0.009300886653363705, -0.0012501750607043505, 0.14612135291099548, -0.017447596415877342, -0.003124968148767948, 0.10153692215681076, -0.0037476997822523117, 0.07193698734045029, 0.0876287892460823, -0.005015614442527294, -0.01397298276424408, 0.1399984061717987, -0.11333011835813522, 0.10244922339916229, 0.1510549634695053, -0.018801162019371986, -0.05093890056014061, -0.014044680632650852, 0.019733866676688194, -0.008319247514009476, 0.03310481086373329, -0.08636397123336792, -0.1024097427725792, 0.007476715371012688, -0.06261133402585983, 0.09766097366809845, -0.10490398108959198, -0.0827425941824913, -0.05013272911310196, -0.009900981560349464, -0.05648289620876312, 0.07076855003833771, 0.07967428863048553, -0.008831257000565529, -0.0446072481572628, -0.09401240199804306, -0.004498292226344347, 0.1222902163863182, -0.07166965305805206, -0.0828961730003357 ]
null
null
transformers
# ukr-t5-small A compact T5-small model fine-tuned for Ukrainian language tasks, with base English understanding. ## Model Description * **Base Model:** mT5-small * **Fine-tuning Data:** Leipzig Corpora Collection (English & Ukrainian news from 2023) * **Tasks:** * Text summarization (Ukrainian) * Text generation (Ukrainian) * Other Ukrainian-centric NLP tasks ## Technical Details * **Model Size:** 300 MB * **Framework:** Transformers (Hugging Face) ## Usage **Installation** ```bash pip install transformers ``` **Loading the Model** ```python from transformers import AutoTokenizer, AutoModelForSeq2SeqLM tokenizer = AutoTokenizer.from_pretrained("path/to/ukr-t5-small") model = AutoModelForSeq2SeqLM.from_pretrained("path/to/ukr-t5-small") ``` **Example: Machine Translation** ```python text = "(Text in Ukrainian here)" # Tokenize and translate inputs = tokenizer("summarize: " + text, return_tensors="pt", max_length=512, truncation=True) summary_ids = model.generate(inputs["input_ids"], num_beams=4, max_length=128) # Decode output summary = tokenizer.decode(summary_ids[0], skip_special_tokens=True) print(summary) ``` ## Limitations * The model's focus is on Ukrainian text processing, so performance on purely English tasks may be below that of general T5-small models. * Further fine-tuning may be required for optimal results on specific NLP tasks. ## Dataset Credits This model was fine-tuned on the Leipzig Corpora Collection (specify if there's a particular subset within the collection that you used). For full licensing and usage information of the original dataset, please refer to [Leipzig Corpora Collection website](https://wortschatz.uni-leipzig.de/en/download) ## Ethical Considerations * NLP models can reflect biases present in their training data. Be mindful of this when using this model for applications that have real-world impact. * It's important to test this model thoroughly across a variety of Ukrainian language samples to evaluate its reliability and fairness.
{"language": ["uk", "en"], "license": "apache-2.0"}
text2text-generation
d0p3/ukr-t5-small
[ "transformers", "safetensors", "t5", "text2text-generation", "uk", "en", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T12:32:41+00:00
[]
[ "uk", "en" ]
TAGS #transformers #safetensors #t5 #text2text-generation #uk #en #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# ukr-t5-small A compact T5-small model fine-tuned for Ukrainian language tasks, with base English understanding. ## Model Description * Base Model: mT5-small * Fine-tuning Data: Leipzig Corpora Collection (English & Ukrainian news from 2023) * Tasks: * Text summarization (Ukrainian) * Text generation (Ukrainian) * Other Ukrainian-centric NLP tasks ## Technical Details * Model Size: 300 MB * Framework: Transformers (Hugging Face) ## Usage Installation Loading the Model Example: Machine Translation ## Limitations * The model's focus is on Ukrainian text processing, so performance on purely English tasks may be below that of general T5-small models. * Further fine-tuning may be required for optimal results on specific NLP tasks. ## Dataset Credits This model was fine-tuned on the Leipzig Corpora Collection (specify if there's a particular subset within the collection that you used). For full licensing and usage information of the original dataset, please refer to Leipzig Corpora Collection website ## Ethical Considerations * NLP models can reflect biases present in their training data. Be mindful of this when using this model for applications that have real-world impact. * It's important to test this model thoroughly across a variety of Ukrainian language samples to evaluate its reliability and fairness.
[ "# ukr-t5-small\n\nA compact T5-small model fine-tuned for Ukrainian language tasks, with base English understanding.", "## Model Description\n\n* Base Model: mT5-small \n* Fine-tuning Data: Leipzig Corpora Collection (English & Ukrainian news from 2023)\n* Tasks: \n * Text summarization (Ukrainian)\n * Text generation (Ukrainian)\n * Other Ukrainian-centric NLP tasks", "## Technical Details\n* Model Size: 300 MB \n* Framework: Transformers (Hugging Face)", "## Usage \n\nInstallation\n\n\n\nLoading the Model\n\n\n\nExample: Machine Translation", "## Limitations\n\n* The model's focus is on Ukrainian text processing, so performance on purely English tasks may be below that of general T5-small models.\n* Further fine-tuning may be required for optimal results on specific NLP tasks.", "## Dataset Credits\n\nThis model was fine-tuned on the Leipzig Corpora Collection (specify if there's a particular subset within the collection that you used). For full licensing and usage information of the original dataset, please refer to Leipzig Corpora Collection website", "## Ethical Considerations\n\n* NLP models can reflect biases present in their training data. Be mindful of this when using this model for applications that have real-world impact.\n* It's important to test this model thoroughly across a variety of Ukrainian language samples to evaluate its reliability and fairness." ]
[ "TAGS\n#transformers #safetensors #t5 #text2text-generation #uk #en #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# ukr-t5-small\n\nA compact T5-small model fine-tuned for Ukrainian language tasks, with base English understanding.", "## Model Description\n\n* Base Model: mT5-small \n* Fine-tuning Data: Leipzig Corpora Collection (English & Ukrainian news from 2023)\n* Tasks: \n * Text summarization (Ukrainian)\n * Text generation (Ukrainian)\n * Other Ukrainian-centric NLP tasks", "## Technical Details\n* Model Size: 300 MB \n* Framework: Transformers (Hugging Face)", "## Usage \n\nInstallation\n\n\n\nLoading the Model\n\n\n\nExample: Machine Translation", "## Limitations\n\n* The model's focus is on Ukrainian text processing, so performance on purely English tasks may be below that of general T5-small models.\n* Further fine-tuning may be required for optimal results on specific NLP tasks.", "## Dataset Credits\n\nThis model was fine-tuned on the Leipzig Corpora Collection (specify if there's a particular subset within the collection that you used). For full licensing and usage information of the original dataset, please refer to Leipzig Corpora Collection website", "## Ethical Considerations\n\n* NLP models can reflect biases present in their training data. Be mindful of this when using this model for applications that have real-world impact.\n* It's important to test this model thoroughly across a variety of Ukrainian language samples to evaluate its reliability and fairness." ]
[ 61, 30, 66, 21, 12, 57, 56, 69 ]
[ "passage: TAGS\n#transformers #safetensors #t5 #text2text-generation #uk #en #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# ukr-t5-small\n\nA compact T5-small model fine-tuned for Ukrainian language tasks, with base English understanding.## Model Description\n\n* Base Model: mT5-small \n* Fine-tuning Data: Leipzig Corpora Collection (English & Ukrainian news from 2023)\n* Tasks: \n * Text summarization (Ukrainian)\n * Text generation (Ukrainian)\n * Other Ukrainian-centric NLP tasks## Technical Details\n* Model Size: 300 MB \n* Framework: Transformers (Hugging Face)## Usage \n\nInstallation\n\n\n\nLoading the Model\n\n\n\nExample: Machine Translation## Limitations\n\n* The model's focus is on Ukrainian text processing, so performance on purely English tasks may be below that of general T5-small models.\n* Further fine-tuning may be required for optimal results on specific NLP tasks.## Dataset Credits\n\nThis model was fine-tuned on the Leipzig Corpora Collection (specify if there's a particular subset within the collection that you used). For full licensing and usage information of the original dataset, please refer to Leipzig Corpora Collection website## Ethical Considerations\n\n* NLP models can reflect biases present in their training data. Be mindful of this when using this model for applications that have real-world impact.\n* It's important to test this model thoroughly across a variety of Ukrainian language samples to evaluate its reliability and fairness." ]
[ -0.04810088500380516, -0.0018965151393786073, -0.0018024018500000238, 0.024495404213666916, 0.13056443631649017, 0.0319400355219841, 0.12555581331253052, 0.06213844195008278, 0.011751469224691391, 0.04283849522471428, -0.05197909101843834, -0.07440099865198135, 0.027377033606171608, 0.041328251361846924, 0.11798960715532303, -0.24818813800811768, 0.04244768247008324, -0.09371974319219589, -0.044583551585674286, 0.028527015820145607, 0.15373726189136505, -0.08349137008190155, 0.03770899027585983, -0.03322811424732208, -0.016515720635652542, 0.06811641156673431, 0.03516490012407303, -0.03137114271521568, 0.11755725741386414, 0.1243203654885292, 0.06725147366523743, -0.05308849364519119, 0.04625359922647476, -0.2064218968153, 0.010294810868799686, 0.06693752855062485, -0.044976573437452316, -0.012992888689041138, 0.10172310471534729, -0.035340454429388046, 0.29343631863594055, -0.1464027762413025, 0.0333855077624321, 0.07551375031471252, -0.05855754017829895, -0.08497510850429535, -0.03520318493247032, 0.12092629075050354, 0.10424381494522095, 0.1183830201625824, -0.061499349772930145, 0.14494210481643677, -0.06472930312156677, 0.09671495854854584, 0.07130663096904755, -0.22388997673988342, -0.0117331612855196, 0.08602230995893478, -0.05311678349971771, 0.11744126677513123, -0.052133958786726, 0.03863001987338066, 0.036204393953084946, 0.0021793365012854338, -0.042298462241888046, -0.005897939670830965, -0.14354808628559113, 0.012216449715197086, -0.14868241548538208, -0.0886848121881485, 0.19955943524837494, -0.01358867809176445, -0.0055900146253407, -0.18406309187412262, -0.026122767478227615, 0.005119697656482458, -0.02652171067893505, -0.029999183490872383, 0.0035790440160781145, -0.01534953061491251, 0.04665691405534744, -0.10867732018232346, -0.13937662541866302, -0.01937033422291279, -0.07829833775758743, 0.14705638587474823, 0.039461519569158554, 0.011246394366025925, 0.04207208752632141, 0.08871415257453918, -0.07618825137615204, -0.03475677967071533, -0.007984047755599022, -0.014292580075562, -0.19467680156230927, -0.063661590218544, -0.047188762575387955, -0.10504188388586044, 0.009371230378746986, 0.10563575476408005, -0.09357306361198425, -0.02199588716030121, 0.029534023255109787, 0.012294085696339607, 0.06838905811309814, 0.09014014154672623, -0.09115000069141388, -0.027020419016480446, -0.04175885021686554, -0.1791267991065979, 0.004997613374143839, 0.012898418121039867, -0.0991087481379509, -0.045266300439834595, -0.002930950140580535, 0.13033953309059143, 0.01445513404905796, 0.021612534299492836, 0.0020131871569901705, -0.06948776543140411, 0.13496527075767517, -0.11338222771883011, -0.03350323811173439, -0.06315785646438599, 0.03338582441210747, -0.0034540018532425165, -0.03855631873011589, 0.026877816766500473, -0.12381728738546371, 0.06801649183034897, -0.028226062655448914, -0.03316917270421982, -0.03553412854671478, -0.11769561469554901, 0.03400709852576256, 0.02046593837440014, -0.0050123766995966434, -0.14249897003173828, -0.10568105429410934, 0.017716430127620697, 0.013469520956277847, -0.016503430902957916, 0.01575179398059845, 0.04489681497216225, -0.02445933409035206, -0.004546450451016426, 0.009011809714138508, -0.012160902842879295, -0.07803022116422653, -0.007000281009823084, -0.07757706195116043, -0.02922637388110161, -0.030049413442611694, -0.006614584010094404, -0.1551889330148697, -0.07795760780572891, -0.1784324198961258, 0.06473720073699951, -0.08687048405408859, 0.027924224734306335, -0.09294747561216354, -0.007419224362820387, -0.008745413273572922, 0.035676393657922745, -0.05510716140270233, 0.06676653772592545, -0.11955824494361877, -0.055573686957359314, 0.0936400517821312, -0.19426333904266357, -0.03937222436070442, 0.1401977241039276, 0.013527186587452888, 0.04428812488913536, 0.16821540892124176, 0.14069652557373047, 0.1408834159374237, -0.1109655499458313, -0.08854055404663086, 0.00009510794916423038, -0.10248172283172607, 0.09143565595149994, 0.09024468064308167, -0.06853017956018448, 0.04084981977939606, 0.021703941747546196, -0.10533018410205841, 0.010457831434905529, 0.003938375972211361, -0.08767209202051163, -0.0266428142786026, -0.02493554726243019, 0.05917249992489815, 0.06424731761217117, -0.015804579481482506, -0.06844466179609299, -0.11716458201408386, 0.11878272145986557, 0.10422881692647934, -0.012164602987468243, 0.07630795985460281, -0.07061758637428284, -0.014613931998610497, 0.030062690377235413, 0.007741009350866079, -0.18004803359508514, -0.15799759328365326, 0.023936545476317406, -0.08687419444322586, 0.1085033118724823, 0.07700059562921524, 0.03519761934876442, 0.07014337182044983, -0.11345108598470688, 0.038595035672187805, -0.060416724532842636, -0.00020242860773578286, -0.028693847358226776, -0.2090417891740799, -0.025318389758467674, -0.04068813472986221, 0.049742814153432846, -0.07231646031141281, -0.015548698604106903, 0.05549023300409317, 0.1570604294538498, 0.05476820841431618, -0.006288198288530111, -0.05640867352485657, 0.03956322744488716, -0.042316045612096786, -0.025742320343852043, 0.016313914209604263, -0.02894146926701069, -0.03584469109773636, 0.08243095129728317, -0.09952466934919357, -0.043799884617328644, 0.0887092873454094, 0.1337108314037323, -0.09151743352413177, 0.006829473190009594, -0.036331452429294586, -0.0010297390399500728, -0.04836536943912506, -0.10936836153268814, 0.1309056133031845, -0.007729930803179741, -0.014746326021850109, -0.08392378687858582, -0.03363515064120293, -0.009346705861389637, 0.02269108034670353, -0.03665554150938988, 0.09336113184690475, 0.02553076483309269, -0.2207496464252472, 0.06804537028074265, -0.05667947977781296, 0.03355149179697037, 0.24894925951957703, -0.045988064259290695, -0.1267215609550476, 0.026027124375104904, -0.005306108854711056, 0.029430538415908813, 0.21664635837078094, -0.1016845703125, 0.02920183725655079, 0.05208809673786163, 0.021767599508166313, 0.03186864033341408, -0.024499278515577316, 0.0287019070237875, 0.012590385973453522, -0.09429898858070374, -0.012392109259963036, 0.02722434513270855, -0.011634956113994122, 0.10262348502874374, -0.01760992407798767, 0.14231166243553162, -0.05559065565466881, -0.0356978215277195, -0.13407480716705322, 0.10195106267929077, -0.09996062517166138, -0.16478894650936127, -0.1319003403186798, 0.1274929791688919, -0.06474266946315765, 0.012908094562590122, 0.07954489439725876, -0.1146407201886177, -0.10147057473659515, -0.0919833555817604, 0.17160363495349884, 0.030078528448939323, -0.06067923083901405, -0.20217609405517578, 0.08072097599506378, -0.045729074627161026, -0.13815483450889587, 0.01251902710646391, -0.0007780687301419675, -0.07440140843391418, -0.053673285990953445, -0.016973251476883888, 0.10658258199691772, 0.1476222276687622, -0.016235539689660072, -0.012343091890215874, -0.08960318565368652, 0.0836290717124939, -0.15576833486557007, 0.05405522510409355, 0.1160648837685585, 0.038858916610479355, -0.025622062385082245, 0.18328429758548737, -0.001991098280996084, 0.0041695209220051765, 0.061849310994148254, 0.09802465885877609, -0.04584044963121414, -0.22429662942886353, -0.20969367027282715, -0.08076716214418411, 0.052329957485198975, 0.013955585658550262, 0.04818928241729736, 0.034193865954875946, -0.025030072778463364, -0.15697911381721497, -0.11350050568580627, 0.039032481610774994, 0.09639929980039597, 0.08698056638240814, 0.036105863749980927, 0.04807894676923752, -0.09815941005945206, -0.022478828206658363, 0.14369986951351166, -0.09760290384292603, 0.23826204240322113, 0.0007918472401797771, 0.17244257032871246, 0.03714975342154503, 0.057314708828926086, -0.019036106765270233, 0.06671395152807236, 0.04865175858139992, 0.02110358513891697, 0.03969484567642212, -0.10746407508850098, 0.06647796183824539, 0.08742176741361618, 0.07136515527963638, -0.08901068568229675, -0.051530107855796814, -0.07151191681623459, 0.0944770872592926, 0.17733804881572723, 0.0440092533826828, -0.07147812843322754, -0.15340836346149445, 0.0348447822034359, -0.07573208957910538, -0.026326173916459084, -0.018572203814983368, 0.10963023453950882, -0.16234293580055237, 0.028458086773753166, 0.0104813352227211, 0.07318562269210815, 0.03849881887435913, -0.009393083862960339, 0.0003585783124435693, 0.0077188890427351, -0.021067503839731216, 0.14573359489440918, -0.11698955297470093, 0.17767509818077087, 0.006807604338973761, 0.0924740731716156, -0.09425397962331772, -0.0036392880138009787, 0.021278806030750275, 0.06477111577987671, 0.12106861174106598, 0.031969159841537476, -0.14648449420928955, -0.05402050539851189, -0.06096973270177841, 0.026835324242711067, 0.06469079852104187, -0.09816357493400574, 0.07187383621931076, -0.04687020555138588, -0.00268483511172235, 0.00037195850745774806, -0.024411413818597794, -0.22562631964683533, -0.160389244556427, 0.022803768515586853, -0.09199433773756027, -0.01129532977938652, -0.030062586069107056, -0.06462334841489792, 0.044536709785461426, 0.13051821291446686, -0.04717421904206276, -0.12221790105104446, -0.12417057156562805, -0.0527709499001503, 0.1577167958021164, -0.0830480083823204, 0.0656612366437912, 0.008897085674107075, 0.16804496943950653, -0.032755084335803986, -0.09800825268030167, 0.04147130250930786, -0.0596284419298172, -0.141724094748497, 0.0485120452940464, 0.13673481345176697, 0.08242063224315643, 0.08199901878833771, -0.005894002038985491, 0.04114913567900658, 0.05554511770606041, -0.10116580128669739, -0.06243910640478134, 0.03432668372988701, 0.03962986171245575, 0.13016030192375183, -0.07929787784814835, -0.1410352885723114, -0.11361756175756454, -0.06849513202905655, 0.11634457111358643, 0.09028217941522598, -0.09050507098436356, 0.16384609043598175, 0.07897302508354187, -0.0801033079624176, -0.23479987680912018, -0.003577153431251645, 0.0016773787792772055, 0.03783469647169113, 0.09087248891592026, -0.11371010541915894, 0.11834118515253067, 0.0094444639980793, -0.0007164690759964287, 0.019674893468618393, -0.24572238326072693, -0.18588480353355408, 0.05616021901369095, 0.030942073091864586, -0.114772729575634, -0.07365375757217407, 0.008489982225000858, -0.06258007884025574, -0.08576973527669907, 0.0852990373969078, -0.15692126750946045, 0.07808820903301239, 0.04197569936513901, -0.02045995183289051, 0.040169455111026764, -0.005847939755767584, 0.09913758188486099, 0.06208013743162155, 0.029640696942806244, -0.08877457678318024, 0.0437355674803257, 0.1829473227262497, 0.0024064488243311644, 0.12118739634752274, -0.008512808009982109, 0.05363617092370987, -0.08486340939998627, -0.06903295964002609, -0.02674802765250206, 0.11929323524236679, -0.053403090685606, -0.11367294192314148, -0.14280524849891663, 0.13103175163269043, 0.05797020345926285, -0.030741265043616295, 0.10458750277757645, -0.07526424527168274, -0.007615397218614817, -0.0061239395290613174, 0.180453360080719, 0.011550601571798325, -0.08722914755344391, 0.03429253399372101, -0.04413268715143204, 0.04521723464131355, -0.10634653270244598, 0.05415203422307968, 0.084766685962677, 0.02387961558997631, 0.003406860865652561, -0.02680649608373642, -0.10527743399143219, -0.03191109001636505, 0.0792657732963562, -0.08457222580909729, -0.21893925964832306, -0.04922029376029968, -0.020782282575964928, -0.051630329340696335, -0.01774473302066326, 0.1031695231795311, -0.08636412024497986, 0.016057675704360008, 0.009961153380572796, -0.0012179466430097818, -0.02686757780611515, 0.10162957012653351, -0.010917361825704575, 0.02551652491092682, -0.05067510902881622, 0.12852618098258972, 0.07227245718240738, -0.09830236434936523, 0.003614999121055007, 0.05325411260128021, -0.12282750755548477, -0.07896869629621506, 0.06714383512735367, 0.05593182519078255, -0.15245644748210907, -0.07162772864103317, -0.002181719057261944, -0.11262614279985428, -0.018246693536639214, 0.10782373696565628, 0.045465998351573944, 0.023115240037441254, -0.024247897788882256, 0.03526685759425163, -0.0702069103717804, 0.11250770837068558, 0.08692143112421036, 0.001644957927055657, -0.09078239649534225, 0.15363703668117523, 0.03238743543624878, -0.01357012614607811, -0.04754578322172165, 0.008389557711780071, -0.05224240571260452, -0.008242184296250343, -0.12542766332626343, -0.0518583245575428, -0.08540323376655579, -0.027212928980588913, -0.01797197200357914, -0.06092723086476326, -0.03701827675104141, 0.03008558787405491, -0.023385435342788696, -0.02007337473332882, -0.04568607732653618, 0.09549038112163544, -0.10639148950576782, 0.0393265075981617, 0.04160041734576225, -0.03722415491938591, 0.11335866153240204, 0.03779022395610809, 0.02227104641497135, 0.15184535086154938, -0.05860771983861923, 0.06664660573005676, -0.06280698627233505, 0.049240171909332275, -0.022816013544797897, -0.08140930533409119, -0.011454329825937748, 0.008566433563828468, 0.05286303535103798, 0.04961157217621803, -0.008623365312814713, -0.08607542514801025, 0.09633505344390869, -0.0205586776137352, -0.053138092160224915, -0.033018141984939575, 0.028608528897166252, 0.04251204803586006, 0.021353650838136673, 0.1503787338733673, -0.09322181344032288, 0.018435906618833542, -0.09885996580123901, 0.008019576780498028, -0.008463992737233639, 0.04636959359049797, -0.0894748717546463, -0.047744426876306534, 0.012977615930140018, -0.03659294173121452, 0.17659923434257507, 0.02551710419356823, 0.0944482684135437, 0.10424361377954483, 0.0525217205286026, 0.10089246928691864, 0.03087863326072693, 0.16230322420597076, -0.0005780525389127433, 0.018558112904429436, -0.057916611433029175, 0.001988873817026615, -0.05813644453883171, -0.0658312737941742, 0.21987731754779816, 0.16474032402038574, 0.05291102081537247, 0.021212048828601837, 0.04048284888267517, -0.09670627117156982, -0.05159469321370125, -0.053408488631248474, 0.027581697329878807, -0.00022042890486773103, -0.07688072323799133, 0.05960351973772049, 0.2575378715991974, -0.20543073117733002, 0.14154702425003052, 0.022417884320020676, -0.068576380610466, -0.043768756091594696, -0.13378307223320007, -0.021508751437067986, -0.09096800535917282, -0.017030933871865273, -0.1226392388343811, -0.010665254667401314, 0.10230783373117447, 0.023364901542663574, -0.06166374683380127, 0.14006273448467255, -0.09057914465665817, -0.09508542716503143, 0.02168714627623558, 0.02361292578279972, 0.09451401978731155, 0.05467817187309265, 0.01308513805270195, 0.0024138193111866713, -0.0004921986255794764, 0.002005290938541293, 0.0814678743481636, -0.032277222722768784, 0.003865891834720969, -0.02412286400794983, -0.01682942733168602, 0.016649795696139336, 0.037063900381326675, 0.03136523440480232, 0.21277247369289398, 0.08282353729009628, -0.035142771899700165, -0.02433840185403824, 0.2799392342567444, -0.018064571544528008, -0.21150003373622894, -0.12587863206863403, 0.14040957391262054, 0.0695490837097168, 0.032553017139434814, 0.03677145391702652, -0.13373975455760956, -0.013973632827401161, 0.13500477373600006, 0.18184705078601837, 0.0770057737827301, 0.0022594279143959284, -0.03988688439130783, 0.0025623594410717487, -0.0035129745956510305, 0.096146360039711, -0.08919459581375122, 0.2985976040363312, -0.04949330911040306, 0.03985235095024109, -0.026646055281162262, 0.03377971053123474, -0.12093713134527206, 0.1081886813044548, -0.0015288727590814233, -0.08517076820135117, -0.012835780158638954, 0.130763441324234, -0.05988257750868797, -0.09720835089683533, -0.005684524308890104, -0.03443218395113945, -0.07482155412435532, 0.030845507979393005, 0.06450875103473663, 0.04822855442762375, 0.07748622447252274, -0.007571315858513117, 0.015298893675208092, 0.051500968635082245, 0.08547750115394592, -0.12018940597772598, -0.12013526260852814, 0.09900114685297012, 0.02961706556379795, 0.2239607572555542, -0.022099899128079414, 0.10916712135076523, 0.08025848120450974, 0.006203794851899147, -0.06821268051862717, 0.019024543464183807, 0.03235546499490738, -0.09194248169660568, 0.08185098320245743, 0.06851841509342194, -0.040355924516916275, 0.10245853662490845, 0.08722427487373352, -0.10930544883012772, 0.06767769902944565, 0.13298989832401276, 0.05725053697824478, -0.060331687331199646, 0.13264530897140503, -0.15915697813034058, 0.11084400117397308, 0.1342339664697647, -0.012959875166416168, -0.0024642522912472486, -0.07549329847097397, 0.04432506486773491, 0.05613447353243828, 0.07607658207416534, -0.0018356264336034656, -0.13212202489376068, -0.0009008980123326182, -0.0038189615588635206, 0.008087338879704475, -0.15053999423980713, -0.07586788386106491, 0.07981850206851959, 0.014782064594328403, -0.023574180901050568, 0.11191423237323761, 0.1381177008152008, -0.03802836686372757, 0.01193989533931017, -0.13922519981861115, 0.03583073616027832, 0.0834316611289978, -0.10492835938930511, -0.09023509919643402 ]
null
null
null
# **Q-Learning** Agent playing1 **FrozenLake-v1** This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** . ## Usage ```python model = load_from_hub(repo_id="0fluff0/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
{"tags": ["FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "q-FrozenLake-v1-4x4-noSlippery", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "FrozenLake-v1-4x4-no_slippery", "type": "FrozenLake-v1-4x4-no_slippery"}, "metrics": [{"type": "mean_reward", "value": "1.00 +/- 0.00", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
0fluff0/q-FrozenLake-v1-4x4-noSlippery
[ "FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
2024-02-10T12:39:10+00:00
[]
[]
TAGS #FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
# Q-Learning Agent playing1 FrozenLake-v1 This is a trained model of a Q-Learning agent playing FrozenLake-v1 . ## Usage
[ "# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage" ]
[ "TAGS\n#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n", "# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage" ]
[ 40, 39 ]
[ "passage: TAGS\n#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage" ]
[ 0.04578453302383423, -0.08074592798948288, -0.00430759321898222, 0.10720831900835037, 0.05034215748310089, -0.040469273924827576, 0.11997015029191971, 0.018999949097633362, 0.20601962506771088, -0.010012076236307621, 0.1455274522304535, 0.007022971753031015, -0.006192410364747047, 0.1867983490228653, 0.04572829231619835, -0.26324528455734253, 0.01831899583339691, -0.09495259821414948, -0.07281816750764847, 0.11870454251766205, 0.05470194295048714, -0.01901467889547348, -0.0007633853238075972, 0.056141503155231476, -0.0673527717590332, 0.0007737681735306978, 0.031996939331293106, -0.012976245954632759, 0.19804789125919342, -0.02254498563706875, 0.06641989201307297, 0.054705578833818436, 0.0758768692612648, -0.1998077929019928, 0.0358855277299881, -0.04215473681688309, -0.09439758956432343, -0.03934839740395546, -0.018780618906021118, 0.05878105387091637, 0.053356342017650604, 0.03858819976449013, 0.058354366570711136, 0.09384993463754654, -0.0773480236530304, 0.04328357055783272, 0.04280758649110794, 0.024811049923300743, 0.04589218273758888, -0.0237203948199749, -0.027002155780792236, 0.08246652781963348, -0.22182892262935638, 0.10318073630332947, -0.010159241035580635, -0.5270710587501526, -0.00633762264624238, 0.24088262021541595, 0.11517096310853958, 0.05707438662648201, -0.06903956830501556, 0.10566288232803345, 0.03913382440805435, -0.007209456991404295, 0.03210983797907829, 0.02150118350982666, 0.12817370891571045, 0.06009242683649063, -0.09581366181373596, 0.040699947625398636, 0.13722525537014008, 0.012822695076465607, 0.020306183025240898, -0.08888901025056839, 0.0410032719373703, -0.03461858257651329, -0.007679527159780264, -0.09758518636226654, 0.05478060990571976, 0.012466507963836193, -0.0934976264834404, -0.09247440844774246, -0.04236573353409767, -0.06708304584026337, 0.11252415925264359, 0.046419668942689896, -0.0874939113855362, 0.03884070739150047, -0.06760413944721222, 0.05918780341744423, -0.16863860189914703, 0.02074250765144825, -0.06627868115901947, -0.09376336634159088, -0.11799788475036621, -0.01683047041296959, -0.07946427166461945, 0.009092256426811218, 0.056664444506168365, 0.1447116881608963, 0.22076484560966492, 0.06690320372581482, 0.09728849679231644, 0.07456006109714508, 0.06531001627445221, 0.1538129299879074, 0.10918238013982773, 0.019075315445661545, -0.015266558155417442, 0.0948706716299057, -0.06445580720901489, -0.1351388692855835, -0.15579092502593994, 0.005488025024533272, 0.0983937531709671, 0.08871900290250778, -0.044080477207899094, -0.006702381651848555, -0.024641724303364754, 0.08566431701183319, -0.11314457654953003, -0.024612564593553543, -0.002267979085445404, 0.06882024556398392, -0.024801667779684067, 0.020378148183226585, -0.06242705136537552, 0.12715265154838562, 0.04222423583269119, -0.059924717992544174, -0.055308472365140915, -0.03053177334368229, -0.014276440255343914, -0.027539284899830818, 0.02446848154067993, -0.07659092545509338, 0.04767750948667526, -0.16766095161437988, -0.042871296405792236, -0.04784649610519409, 0.025697942823171616, -0.03907240927219391, -0.13557587563991547, -0.17699143290519714, -0.048906855285167694, -0.022438718006014824, 0.03549358621239662, -0.038111843168735504, 0.006551501806825399, -0.006318534724414349, -0.1583600640296936, 0.09783563017845154, 0.09784027189016342, -0.03643378987908363, -0.02749447710812092, 0.056263517588377, -0.07194498926401138, 0.1561182290315628, -0.21054518222808838, -0.054014235734939575, -0.044764336198568344, -0.06595750898122787, 0.19673264026641846, 0.012690845876932144, -0.01202624011784792, 0.19873127341270447, -0.29073721170425415, -0.06078760325908661, 0.12533614039421082, -0.07834373414516449, -0.0936407670378685, 0.06941844522953033, -0.04206686094403267, 0.023345354944467545, 0.046047765761613846, 0.36345911026000977, -0.02069227211177349, -0.16197136044502258, -0.021782705560326576, 0.13971707224845886, -0.1184760183095932, 0.059895481914281845, 0.04240793362259865, 0.12543781101703644, -0.04250509291887283, -0.018672896549105644, -0.09023164212703705, 0.05999075248837471, -0.05241934582591057, -0.09016361832618713, -0.03393383324146271, -0.07645075023174286, 0.13294468820095062, -0.0629684180021286, 0.05601520463824272, -0.03255095332860947, -0.07133250683546066, -0.050324998795986176, -0.016492370516061783, 0.04460815340280533, 0.05951254442334175, -0.12794871628284454, 0.11029167473316193, 0.13025271892547607, -0.0006193425506353378, -0.07498852163553238, -0.17872096598148346, 0.003240168560296297, 0.009576505981385708, 0.039837226271629333, 0.17141658067703247, 0.12209978699684143, 0.033295199275016785, 0.008770671673119068, -0.06389404833316803, -0.18276847898960114, 0.058129217475652695, -0.056212130934000015, -0.14230976998806, -0.052409034222364426, -0.0728459507226944, 0.017381802201271057, -0.0859743058681488, -0.017379917204380035, 0.021926190704107285, 0.006908397190272808, 0.02990424446761608, -0.026645656675100327, -0.049561817198991776, 0.021254703402519226, 0.06490101665258408, -0.0037617047782987356, 0.12023693323135376, 0.008277264423668385, -0.18308481574058533, 0.07930773496627808, 0.08478537946939468, 0.09196605533361435, 0.013250201940536499, 0.02685922384262085, -0.021522263064980507, -0.08061408251523972, -0.054420311003923416, 0.02957955375313759, 0.11417073011398315, 0.1317172348499298, 0.2361993044614792, 0.08753683418035507, 0.04697408527135849, -0.02164587564766407, -0.016415923833847046, 0.002810494042932987, -0.06318057328462601, -0.029935607686638832, 0.10614971816539764, 0.05865858122706413, -0.067733034491539, -0.04576427489519119, 0.09590928256511688, 0.02732124738395214, 0.21205885708332062, -0.03342745825648308, 0.01286078616976738, -0.10957037657499313, -0.06550975888967514, -0.031982194632291794, 0.09201868623495102, 0.09498392790555954, 0.009755023755133152, -0.022056059911847115, -0.04259001836180687, 0.0012916827108711004, -0.1334889680147171, -0.10375088453292847, 0.026475343853235245, 0.013400445692241192, -0.11206940561532974, 0.11674030870199203, -0.11352457851171494, 0.039504457265138626, 0.06024791672825813, -0.13837239146232605, 0.04428480193018913, -0.029713207855820656, -0.07886212319135666, 0.16866780817508698, -0.11075661331415176, -0.094340018928051, -0.08831550180912018, 0.004082420375198126, 0.0075836325995624065, -0.03922267258167267, -0.009283260442316532, -0.19952571392059326, -0.005375816952437162, -0.03544965013861656, 0.013616434298455715, -0.06988783925771713, -0.11287739872932434, -0.010957922786474228, 0.07084179669618607, -0.043388739228248596, -0.07803605496883392, 0.007967432029545307, -0.08923084288835526, -0.10623309016227722, 0.028189711272716522, 0.019765101373195648, -0.022883659228682518, 0.16152891516685486, 0.01816628873348236, 0.05626589432358742, -0.03298520669341087, 0.30665266513824463, -0.038163769990205765, 0.08371731638908386, -0.02993497997522354, -0.07433546334505081, 0.06130730360746384, -0.022327827289700508, 0.06086638569831848, -0.020221687853336334, -0.02362890914082527, 0.0077952733263373375, -0.08579335361719131, -0.18365982174873352, -0.05417544022202492, 0.03724347800016403, 0.195254847407341, 0.031118987128138542, 0.01910330168902874, -0.0488768145442009, -0.010547760874032974, 0.1665220558643341, -0.10005921125411987, 0.04030545800924301, -0.05366240441799164, 0.11506262421607971, -0.08640182018280029, 0.06195629760622978, 0.020486772060394287, 0.04266135022044182, -0.04877188801765442, 0.09486009180545807, 0.0826394334435463, 0.1121082529425621, -0.02206910029053688, 0.046257395297288895, 0.019012698903679848, 0.07383184134960175, 0.11073657125234604, 0.0368414968252182, -0.0729052945971489, 0.001982470043003559, -0.006313489284366369, -0.039427030831575394, 0.11933320760726929, 0.17963355779647827, -0.11991413682699203, -0.05106910318136215, 0.27167606353759766, 0.0031242913100868464, 0.19481229782104492, -0.01315275114029646, 0.043591804802417755, -0.04484925419092178, 0.04572054371237755, -0.05338600277900696, -0.04086209088563919, 0.2094656229019165, 0.08045925945043564, -0.17165091633796692, -0.08549032360315323, -0.05912299454212189, 0.07081323862075806, 0.10728751868009567, 0.0013539529172703624, -0.04156802222132683, 0.0004610282776411623, 0.0014198932331055403, 0.08339415490627289, -0.14520122110843658, 0.11816094070672989, -0.03172019124031067, 0.05612684786319733, 0.017555562779307365, -0.045326150953769684, 0.04264266416430473, 0.07474290579557419, 0.26618310809135437, 0.0904107540845871, -0.040318213403224945, -0.0892091691493988, -0.12260187417268753, 0.010461576282978058, 0.029102616012096405, -0.03534553572535515, 0.0037547778338193893, -0.020087555050849915, 0.0318896509706974, 0.008264793083071709, 0.016230624169111252, -0.08987458795309067, -0.03175399824976921, -0.027736429125070572, -0.023839212954044342, 0.10733365267515182, -0.09495144337415695, -0.1444292515516281, -0.15713949501514435, 0.04191131144762039, -0.0766405463218689, -0.056593164801597595, -0.054507751017808914, -0.05239389091730118, -0.0311186034232378, -0.03773957118391991, 0.09099467098712921, -0.0021037792321294546, 0.14807306230068207, -0.1920108050107956, -0.04220759496092796, 0.051812779158353806, -0.07607918977737427, -0.08729588985443115, 0.03410962224006653, 0.12136995792388916, 0.05116051807999611, 0.11504370719194412, 0.013609255664050579, 0.09567681699991226, 0.0045484392903745174, -0.06713183224201202, 0.15302421152591705, -0.14069625735282898, -0.27875974774360657, -0.03836318850517273, 0.016946332529187202, 0.1615200787782669, -0.05613167956471443, 0.031766023486852646, 0.3335736393928528, 0.27782970666885376, -0.1428707242012024, 0.25916144251823425, 0.019178593531250954, 0.004398873541504145, -0.19130495190620422, -0.10125631093978882, 0.025324683636426926, 0.04740457236766815, 0.12032642960548401, -0.14564448595046997, -0.010732659138739109, -0.04543145373463631, -0.025908485054969788, 0.10386138409376144, -0.12300799041986465, -0.07263197749853134, 0.07765276730060577, 0.039809420704841614, 0.1808302253484726, 0.03932500258088112, 0.0014799144119024277, 0.13626977801322937, 0.06612244248390198, 0.019124457612633705, 0.05216038227081299, 0.08028066903352737, -0.018944554030895233, 0.14207926392555237, 0.05448179319500923, -0.02551644667983055, 0.052681710571050644, -0.0054580713622272015, -0.03219012916088104, 0.015605825930833817, -0.183198019862175, -0.10147556662559509, -0.0561356320977211, -0.10798973590135574, -0.04978342354297638, 0.056853994727134705, -0.12395523488521576, -0.007896827533841133, -0.03841273859143257, 0.03718273714184761, -0.07831971347332001, -0.09360362589359283, -0.036494381725788116, 0.1351792961359024, 0.07210618257522583, 0.04471297934651375, 0.035655103623867035, -0.07390819489955902, 0.07097936421632767, 0.21671734750270844, 0.08159157633781433, 0.028919655829668045, -0.19545674324035645, -0.024042490869760513, -0.0803457647562027, 0.06306298077106476, -0.08856996893882751, -0.016788700595498085, 0.11923003196716309, 0.08616556972265244, 0.05413002520799637, 0.09640096127986908, -0.045083072036504745, 0.021686913445591927, 0.02684609219431877, -0.15131035447120667, -0.18501274287700653, -0.08534606546163559, -0.03519878163933754, 0.11561143398284912, -0.06398691236972809, 0.10897188633680344, -0.13615410029888153, 0.010051886551082134, -0.006060056854039431, 0.02693452313542366, -0.03596206381917, -0.11251141875982285, 0.15348562598228455, 0.11999429017305374, -0.06767056882381439, 0.03127254918217659, -0.09527092427015305, -0.04423454403877258, 0.12686803936958313, -0.013623855076730251, -0.0371493324637413, -0.054547641426324844, -0.03628576174378395, 0.15247689187526703, -0.03436964750289917, 0.008244883269071579, -0.041229065507650375, -0.18217355012893677, 0.0798322781920433, 0.09045056998729706, 0.019827889278531075, -0.031874191015958786, -0.09797266125679016, -0.010231015272438526, -0.0011165260802954435, 0.11730700731277466, -0.10696814209222794, -0.10933240503072739, -0.15144047141075134, 0.06713984161615372, -0.0007159380475059152, 0.18502596020698547, -0.06394898891448975, -0.08904669433832169, -0.12429379671812057, 0.02344517596065998, -0.0027384376153349876, -0.042264558374881744, 0.01618490368127823, 0.07992301136255264, -0.04095321521162987, 0.02075677551329136, -0.06651144474744797, 0.06372585147619247, -0.11786920577287674, 0.09625071287155151, 0.01063506118953228, 0.016993753612041473, -0.0417880080640316, -0.01618220843374729, 0.039470795542001724, -0.057925306260585785, 0.07921463251113892, 0.011758086271584034, 0.0010938759660348296, 0.10196787863969803, -0.0034960443153977394, 0.06409632414579391, -0.05372481048107147, -0.023290161043405533, 0.06578411161899567, -0.05874887853860855, -0.03370826691389084, -0.1573946475982666, -0.0709633082151413, 0.020051732659339905, -0.04775108024477959, 0.002077929675579071, 0.03673801198601723, 0.062159497290849686, -0.06937079131603241, -0.12125655263662338, -0.043812792748212814, -0.028638383373618126, 0.021301284432411194, 0.10829301923513412, -0.07526551932096481, 0.1547859013080597, -0.052787959575653076, -0.00020603960729204118, 0.07437096536159515, 0.04048224538564682, 0.01393822580575943, -0.10422444343566895, -0.04698587954044342, -0.11035211384296417, 0.1502903699874878, -0.007902312092483044, -0.03533121198415756, 0.03719403222203255, -0.11946307867765427, -0.1572723090648651, 0.03418220207095146, 0.10199101269245148, 0.0448341928422451, 0.025807438418269157, 0.027079269289970398, -0.04042419046163559, -0.021270349621772766, -0.07034418731927872, 0.0882953479886055, -0.12085357308387756, -0.09669415652751923, 0.09555385261774063, 0.12178351730108261, -0.0036850625183433294, -0.07441367954015732, 0.11554073542356491, -0.021787192672491074, 0.05525410920381546, -0.02971339225769043, 0.10308072715997696, 0.0796005055308342, -0.12273547053337097, 0.005693064536899328, -0.036891788244247437, -0.0741485133767128, -0.12975730001926422, 0.019545545801520348, -0.061916105449199677, -0.13383042812347412, 0.12179028987884521, -0.09376577287912369, 0.030037038028240204, -0.10506992787122726, 0.021338803693652153, 0.01864001713693142, 0.061665527522563934, -0.10988292098045349, 0.08575301617383957, 0.13424484431743622, -0.043199893087148666, -0.07184189558029175, -0.12455986440181732, -0.05022053420543671, -0.04231856390833855, -0.13957437872886658, -0.11600435525178909, 0.0100301094353199, -0.023418782278895378, -0.05818291753530502, 0.0015462689334526658, -0.03659068048000336, 0.008594646118581295, 0.021907730028033257, 0.04032021388411522, -0.02693161368370056, 0.05134565755724907, -0.057569269090890884, -0.052510857582092285, 0.11489357799291611, 0.04113486409187317, -0.03561042994260788, -0.052359987050294876, 0.12997733056545258, -0.11959461867809296, 0.07662346214056015, -0.020313527435064316, 0.017129231244325638, -0.06435854732990265, 0.17131924629211426, 0.11673715710639954, -0.1367570012807846, -0.005008010193705559, -0.08210669457912445, 0.020409544929862022, 0.023555370047688484, 0.13693512976169586, -0.03411718085408211, -0.0012358218664303422, -0.1580323874950409, 0.018575575202703476, -0.18557456135749817, -0.03716109320521355, 0.04671547934412956, 0.09917585551738739, 0.15293832123279572, -0.0034432117827236652, -0.1263325810432434, 0.10424192249774933, -0.2118520885705948, 0.0907607227563858, 0.05121984705328941, -0.11874113976955414, -0.06765396893024445, -0.06795281916856766, 0.1198519766330719, 0.009196433238685131, 0.2040700763463974, -0.013615905307233334, -0.09132910519838333, -0.07060808688402176, -0.01980910450220108, -0.030524181202054024, 0.09714830666780472, 0.041414931416511536, 0.04653804749250412, 0.12821412086486816, 0.00368314771912992, 0.07533777505159378, 0.060310911387205124, 0.02759413793683052, -0.012300663627684116, 0.04076618701219559, 0.08261215686798096, -0.14588621258735657, -0.1659701019525528, 0.1326720416545868, 0.025149408727884293, 0.11792458593845367, 0.03658788278698921, -0.1549617499113083, 0.06687124073505402, 0.2523096203804016, -0.11147607117891312, 0.02505038119852543, 0.12737524509429932, -0.0366884209215641, 0.0672016367316246, 0.1144871786236763, -0.02633814327418804, -0.05217865854501724, -0.011363590136170387, 0.10233135521411896, 0.028660254552960396, -0.04646271467208862, -0.02340836264193058, -0.03373933956027031, -0.019070526584982872, -0.011738128960132599, -0.0909019410610199, -0.1543993502855301, -0.10471053421497345, -0.16619662940502167, 0.04399140924215317, -0.04626438021659851, 0.13418889045715332, 0.09469578415155411, -0.012723101302981377, 0.04568437114357948, 0.028575526550412178, 0.07275456190109253, 0.07916246354579926, -0.02939477376639843, -0.036159269511699677 ]
null
null
transformers
# Pasta-Sea-7b-128k 8bpw Exllama quant of [Test157t/Pasta-Sea-7b-128k](https://huggingface.co/Test157t/Pasta-Sea-7b-128k) ## Prompt format: Alpaca ``` Below is an instruction that describes a task. Write a response that appropriately completes the request. ### Instruction: {prompt} ### Input: {input} ### Response: ``` ## Contact Kooten on discord [ko-fi.com/kooten](https://ko-fi.com/kooten) if you would like to support me
{"library_name": "transformers", "tags": ["mergekit", "merge"], "base_model": ["Test157t/Kunocchini-7b-128k-test", "Test157t/Pasta-Lake-7b"]}
text-generation
Kooten/Pasta-Sea-7b-128k-8bpw-exl2
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "base_model:Test157t/Kunocchini-7b-128k-test", "base_model:Test157t/Pasta-Lake-7b", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T12:40:31+00:00
[]
[]
TAGS #transformers #safetensors #mistral #text-generation #mergekit #merge #base_model-Test157t/Kunocchini-7b-128k-test #base_model-Test157t/Pasta-Lake-7b #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Pasta-Sea-7b-128k 8bpw Exllama quant of Test157t/Pasta-Sea-7b-128k ## Prompt format: Alpaca ## Contact Kooten on discord URL if you would like to support me
[ "# Pasta-Sea-7b-128k 8bpw\nExllama quant of Test157t/Pasta-Sea-7b-128k", "## Prompt format: Alpaca", "## Contact\nKooten on discord\n\nURL if you would like to support me" ]
[ "TAGS\n#transformers #safetensors #mistral #text-generation #mergekit #merge #base_model-Test157t/Kunocchini-7b-128k-test #base_model-Test157t/Pasta-Lake-7b #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Pasta-Sea-7b-128k 8bpw\nExllama quant of Test157t/Pasta-Sea-7b-128k", "## Prompt format: Alpaca", "## Contact\nKooten on discord\n\nURL if you would like to support me" ]
[ 90, 33, 8, 14 ]
[ "passage: TAGS\n#transformers #safetensors #mistral #text-generation #mergekit #merge #base_model-Test157t/Kunocchini-7b-128k-test #base_model-Test157t/Pasta-Lake-7b #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Pasta-Sea-7b-128k 8bpw\nExllama quant of Test157t/Pasta-Sea-7b-128k## Prompt format: Alpaca## Contact\nKooten on discord\n\nURL if you would like to support me" ]
[ -0.13665421307086945, 0.03328120335936546, -0.0023977274540811777, 0.010874534957110882, 0.030814211815595627, 0.0052447132766246796, 0.1293209344148636, 0.13238875567913055, 0.17011351883411407, 0.04021481052041054, 0.0887300968170166, 0.04026902839541435, 0.046389609575271606, 0.1818971186876297, -0.08776205033063889, -0.10393071174621582, 0.05500900745391846, 0.01404658704996109, 0.042717985808849335, 0.1634928584098816, 0.0969085842370987, -0.06761901080608368, 0.10274248570203781, -0.062132328748703, -0.021953262388706207, 0.08013993501663208, 0.014364074915647507, -0.10090426355600357, 0.09181226789951324, 0.04749181494116783, 0.12797346711158752, 0.07394697517156601, -0.02831091731786728, -0.14941003918647766, 0.06479524821043015, 0.03624352067708969, 0.03407234698534012, 0.02668621391057968, 0.05695432052016258, 0.0017150399507954717, 0.004081965424120426, -0.02448309399187565, -0.035533081740140915, 0.060361724346876144, -0.06962823867797852, 0.04998458921909332, -0.10573034733533859, 0.0719732791185379, 0.120529405772686, 0.1341957449913025, -0.013217432424426079, 0.15357129275798798, -0.06617047637701035, 0.09749434888362885, 0.14698347449302673, -0.1146080419421196, -0.09648855775594711, 0.05112016573548317, 0.07067233324050903, 0.06846624612808228, 0.04147129878401756, 0.06419187784194946, 0.0863381028175354, -0.008742050267755985, -0.18478550016880035, -0.052482862025499344, 0.050813719630241394, -0.04238513112068176, -0.05671472102403641, -0.011776665225625038, 0.21042340993881226, 0.04461514204740524, -0.09291673451662064, 0.0295840036123991, -0.07135430723428726, 0.03232331946492195, 0.005407380405813456, -0.039796773344278336, -0.003772856667637825, 0.005762528162449598, 0.019501008093357086, -0.05365671217441559, -0.10938600450754166, -0.10707023739814758, -0.12866142392158508, 0.24790002405643463, 0.0201522558927536, 0.0026719211600720882, -0.07784321159124374, 0.06411948055028915, -0.17430368065834045, -0.11487966030836105, -0.05008549243211746, -0.053849510848522186, 0.08170753717422485, -0.04373565688729286, 0.0015203009825199842, -0.22011388838291168, 0.15262292325496674, 0.17819836735725403, -0.042776014655828476, 0.06616492569446564, -0.05399386212229729, 0.0423162467777729, -0.027423709630966187, 0.029476987197995186, -0.04516828805208206, -0.08815134316682816, 0.09466202557086945, 0.09018084406852722, 0.04764893651008606, 0.014508018270134926, -0.037799183279275894, -0.10248305648565292, 0.029364200308918953, 0.05711101368069649, 0.10390200465917587, 0.015046438202261925, -0.038057342171669006, -0.0031856561545282602, 0.21537436544895172, -0.11773882806301117, 0.026355713605880737, 0.034302543848752975, -0.060048483312129974, 0.09833764284849167, 0.034449536353349686, 0.036578219383955, 0.011092555709183216, 0.0241576936095953, -0.03031742013990879, -0.00405255239456892, -0.036243986338377, -0.03130914270877838, 0.12183692306280136, -0.013922653160989285, -0.05494014173746109, -0.20599742233753204, -0.19408193230628967, -0.03551280498504639, -0.024614492431282997, -0.06892401725053787, -0.012335079722106457, 0.0023259916342794895, 0.04308543726801872, -0.072316974401474, -0.030211471021175385, -0.02146567404270172, -0.05555466189980507, 0.03227882459759712, 0.0950850397348404, 0.05466103181242943, -0.1229470893740654, 0.010368486866354942, -0.10399842262268066, 0.0612531378865242, -0.18342024087905884, -0.00949668139219284, -0.13043877482414246, 0.11344288289546967, -0.15108323097229004, -0.04202626645565033, -0.1287555992603302, 0.022079484537243843, 0.11850474774837494, 0.22690901160240173, -0.11982563883066177, 0.012596666812896729, 0.11269109696149826, -0.14235429465770721, -0.13466575741767883, 0.1688554584980011, 0.04730069637298584, 0.027225229889154434, 0.12089516967535019, 0.05111377686262131, 0.07504085451364517, -0.07585117965936661, -0.026067044585943222, -0.0326591357588768, -0.022691259160637856, -0.09182322770357132, 0.07059627771377563, -0.031988974660634995, -0.09103037416934967, 0.0393335297703743, 0.0663369745016098, 0.0637718066573143, -0.0178646519780159, -0.053850382566452026, -0.07154927402734756, -0.07728246599435806, 0.03803639113903046, -0.0037227075081318617, 0.012004195712506771, -0.09889236092567444, -0.06341008841991425, -0.10628073662519455, 0.0759982094168663, -0.015201156958937645, 0.06362881511449814, -0.06058485805988312, 0.287831574678421, -0.09417549520730972, -0.009611916728317738, -0.07132205367088318, -0.06629791855812073, 0.00019855807477142662, 0.10399964451789856, 0.000525619019754231, -0.027536677196621895, 0.07831021398305893, -0.06053832173347473, -0.04029129445552826, -0.0160116758197546, 0.13142813742160797, 0.007789533585309982, -0.08192694187164307, -0.05736807733774185, 0.04718055948615074, -0.01532055251300335, 0.06965231150388718, -0.14250268042087555, 0.059018999338150024, 0.22157908976078033, 0.1348382830619812, -0.0705401822924614, 0.014532488770782948, 0.05250991880893707, 0.02490253746509552, -0.05387251824140549, -0.020461561158299446, 0.041060976684093475, -0.07961487770080566, -0.14096669852733612, 0.060473423451185226, -0.016090529039502144, 0.07172251492738724, 0.09303060173988342, 0.009598106145858765, -0.0515720471739769, -0.06786192953586578, -0.036048710346221924, -0.013178345747292042, -0.011936516501009464, -0.00828357134014368, -0.0026407961267977953, 0.03998061269521713, 0.06996636092662811, -0.08937947452068329, 0.03730654716491699, 0.0444033145904541, -0.09492433071136475, -0.037852589040994644, 0.11883959174156189, -0.052986301481723785, -0.16013188660144806, 0.06497196108102798, 0.1571168750524521, -0.09914383292198181, 0.0783109962940216, 0.02951844595372677, -0.04601648449897766, 0.05074698105454445, 0.12867993116378784, -0.00623472360894084, 0.07985411584377289, -0.12544411420822144, 0.06677559018135071, 0.04190349951386452, 0.011342190206050873, 0.040672000497579575, -0.12860429286956787, -0.02208714745938778, 0.004316091071814299, -0.05333343520760536, -0.014520174823701382, 0.0710541307926178, -0.013677054084837437, 0.08441299200057983, 0.04451402276754379, -0.05842328071594238, 0.004750687628984451, 0.00509664136916399, -0.05005841702222824, 0.2154238373041153, -0.019460592418909073, -0.15071730315685272, -0.17197172343730927, 0.0006951401592232287, -0.07194077223539352, -0.020883187651634216, 0.06902487576007843, -0.1431262344121933, -0.010423168540000916, -0.0991576761007309, -0.05608673021197319, 0.10826176404953003, 0.005362760741263628, -0.06191558763384819, -0.0012878001434728503, -0.0012891427613794804, -0.06340242177248001, -0.05454537644982338, -0.00046310125617310405, -0.05067550763487816, 0.08852118998765945, -0.10384859144687653, 0.1296001374721527, 0.1197490319609642, -0.061761826276779175, -0.004838970489799976, 0.0008682790212333202, 0.10168558359146118, -0.07804683595895767, 0.13169026374816895, 0.27480897307395935, 0.011109020560979843, 0.07683577388525009, 0.1249358206987381, -0.004255151376128197, -0.04908610135316849, 0.028442509472370148, 0.022222138941287994, -0.05811520665884018, -0.2748413681983948, -0.047500766813755035, -0.00996600091457367, 0.1258430927991867, -0.02586895041167736, 0.06623683869838715, 0.08208413422107697, 0.15785548090934753, 0.00001760488339641597, -0.12351378053426743, 0.016783373430371284, 0.07424323260784149, 0.17900961637496948, -0.03778042644262314, 0.12741884589195251, -0.051297105848789215, 0.006532344967126846, 0.10130764544010162, 0.14148707687854767, 0.01898559369146824, 0.0003495542041491717, 0.011125564575195312, 0.07141049951314926, 0.08181555569171906, 0.014744450338184834, 0.04165354743599892, 0.03238530084490776, -0.05156029015779495, 0.003227484179660678, -0.08846989274024963, -0.02292628400027752, 0.0368025079369545, -0.05471333861351013, 0.023278700187802315, -0.0015412364155054092, -0.041328564286231995, 0.05230208486318588, 0.08849725872278214, 0.05381673946976662, -0.25164005160331726, -0.05879063159227371, 0.05042262375354767, -0.009271841496229172, -0.036889031529426575, -0.010294855572283268, -0.002315535442903638, -0.07208274304866791, 0.11345282196998596, -0.037114761769771576, 0.11950381845235825, 0.04101841151714325, -0.013971660286188126, 0.03223385289311409, 0.07179389148950577, -0.0508534200489521, 0.07552139461040497, -0.14392222464084625, 0.21450118720531464, -0.0013498460175469518, 0.00661103380843997, -0.03432183712720871, 0.025261618196964264, 0.0015590023249387741, 0.2608036398887634, 0.10744999349117279, 0.011640573851764202, -0.01481802761554718, -0.13241642713546753, -0.16205695271492004, 0.06798456609249115, -0.005781061016023159, 0.061872780323028564, 0.030798031017184258, 0.045152463018894196, -0.01253381185233593, -0.03730057179927826, 0.0551198348402977, -0.2170582264661789, -0.0033239347394555807, 0.021299513056874275, 0.08719155192375183, 0.0618169791996479, -0.07260030508041382, -0.0810660794377327, -0.026095636188983917, 0.22856898605823517, 0.022670039907097816, -0.05643673613667488, -0.10778447240591049, -0.08109968900680542, 0.07943019270896912, -0.0995464101433754, 0.003591127460822463, -0.08613654226064682, -0.030878616496920586, -0.04312630742788315, 0.0034630862064659595, 0.12003906071186066, -0.07197224348783493, -0.11805261671543121, -0.024120545014739037, 0.1251736879348755, -0.03468345105648041, -0.018183425068855286, 0.024383608251810074, -0.029804570600390434, -0.014892772771418095, -0.058336786925792694, -0.0396406352519989, 0.08219783008098602, 0.049141425639390945, 0.04573338106274605, -0.014809626154601574, -0.05977886542677879, -0.11103656888008118, -0.03836871683597565, 0.09869026392698288, 0.2568843364715576, 0.059988293796777725, -0.014858907088637352, 0.1052376851439476, -0.01647442765533924, -0.19256308674812317, 0.058695968240499496, -0.11543876677751541, 0.014111143536865711, -0.08371252566576004, -0.024395423009991646, 0.03388015180826187, 0.16678838431835175, -0.051860325038433075, 0.12690679728984833, -0.19171127676963806, -0.1386745125055313, 0.05096528306603432, 0.09294445067644119, 0.24234741926193237, -0.14458364248275757, -0.0517965666949749, -0.12603239715099335, -0.19701319932937622, -0.08528971672058105, -0.059454191476106644, 0.04186474531888962, -0.05642637610435486, 0.041258953511714935, 0.0032106211874634027, -0.081551693379879, 0.14961043000221252, -0.04399585723876953, -0.031304243952035904, -0.04938340187072754, 0.0015881912549957633, 0.03306136652827263, -0.027677549049258232, 0.11493366956710815, -0.0622757151722908, -0.020213060081005096, -0.15067996084690094, -0.0437653549015522, -0.07218479365110397, 0.1104673445224762, -0.004328192677348852, -0.03213188424706459, -0.02130722999572754, 0.0560351125895977, -0.039404161274433136, -0.010983061045408249, 0.1197676882147789, -0.1417391300201416, 0.1152302473783493, 0.18902713060379028, 0.1556079089641571, -0.09874802082777023, 0.05477451905608177, 0.07102398574352264, -0.09188777953386307, 0.031638793647289276, -0.1541287899017334, -0.004299307242035866, 0.05840069055557251, -0.04263097047805786, -0.011393926106393337, 0.010081164538860321, -0.10955365747213364, 0.011522799730300903, 0.10766485333442688, -0.10673375427722931, -0.05257892608642578, -0.00784590095281601, 0.003517809323966503, -0.07323582470417023, 0.0917762815952301, 0.16444583237171173, -0.06580807268619537, -0.029542183503508568, -0.0033563829492777586, 0.004809269215911627, -0.1129385307431221, 0.16215263307094574, 0.09373586624860764, 0.028900114819407463, -0.010548998601734638, 0.07005487382411957, -0.05081178992986679, -0.11510112881660461, -0.014902004040777683, 0.08064135164022446, -0.07648362219333649, -0.12177703529596329, -0.02889302186667919, 0.029848938807845116, -0.05642601102590561, -0.12168548256158829, -0.16614368557929993, -0.07972141355276108, 0.003679543500766158, 0.15466362237930298, 0.08296897262334824, 0.02049039490520954, 0.05434631183743477, -0.03485041484236717, -0.016390550881624222, 0.06386065483093262, 0.12284856289625168, 0.09271478652954102, -0.11408574879169464, -0.02103041484951973, -0.06974127143621445, 0.0419580303132534, -0.04721837490797043, 0.018245115876197815, -0.10839615017175674, 0.001078633707948029, -0.2122836858034134, -0.018093135207891464, -0.11725445836782455, -0.029910381883382797, -0.050951384007930756, -0.053441595286130905, -0.02318243868649006, 0.004077959805727005, -0.028139246627688408, 0.044455718249082565, -0.03769386187195778, 0.014123651199042797, -0.12000956386327744, 0.015275157988071442, 0.09168336540460587, -0.01530508603900671, 0.01851528510451317, -0.0034263276029378176, -0.043293558061122894, -0.06082375720143318, -0.22802381217479706, 0.08463994413614273, 0.05541694164276123, -0.024396253749728203, -0.006237799767404795, -0.035159435123205185, 0.04106093943119049, 0.06700605154037476, 0.0938132032752037, 0.06695052236318588, 0.09387869387865067, -0.09024116396903992, 0.00687812827527523, 0.013619691133499146, -0.10510896891355515, 0.0006664195097982883, 0.008457845076918602, 0.12669070065021515, 0.04695005342364311, 0.0722053050994873, -0.09482455998659134, 0.05040053650736809, -0.10999844968318939, 0.01790107600390911, 0.030804742127656937, -0.10884221643209457, 0.008007603697478771, -0.06167128309607506, 0.042556051164865494, 0.01427366677671671, 0.05246583744883537, 0.025456387549638748, 0.025533189997076988, 0.05391576886177063, -0.04882831126451492, 0.15108415484428406, 0.037900298833847046, 0.1480136513710022, 0.009349791333079338, 0.00911852065473795, -0.07485264539718628, -0.0026341592893004417, -0.0014792237197980285, -0.10238353163003922, 0.03832506388425827, 0.2126644402742386, -0.016885705292224884, 0.12444541603326797, 0.07162515819072723, 0.10288892686367035, -0.0000380163146473933, -0.16340987384319305, -0.14087674021720886, 0.016942474991083145, -0.03888015076518059, 0.08493034541606903, 0.08432631194591522, -0.1055806428194046, -0.03998804837465286, -0.030152330175042152, -0.06048409640789032, -0.14007431268692017, -0.21421940624713898, -0.09915690124034882, -0.15932932496070862, 0.05008586868643761, -0.02378854900598526, 0.014399201609194279, 0.002313127974048257, 0.01887570694088936, -0.029621845111250877, 0.24343399703502655, -0.0005542339640669525, -0.006480129901319742, 0.09712014347314835, 0.004507651552557945, 0.01832839474081993, 0.10341502726078033, -0.051142752170562744, 0.054869551211595535, -0.07209349423646927, -0.014114512130618095, 0.04696261137723923, -0.006888139992952347, 0.03360467404127121, -0.031059972941875458, -0.1060105636715889, -0.06101664528250694, 0.0569753460586071, 0.07906179875135422, 0.022039396688342094, 0.009256630204617977, 0.043788518756628036, 0.04447973147034645, 0.15533070266246796, -0.0331975556910038, -0.10180777311325073, -0.06080563738942146, 0.18191467225551605, -0.0280098095536232, 0.06415419280529022, -0.05814028158783913, -0.08578812330961227, 0.08899739384651184, 0.18836072087287903, 0.10824446380138397, -0.008491070941090584, -0.007472972851246595, 0.02531496062874794, 0.01716596633195877, -0.03252048045396805, 0.03905612230300903, 0.027120957151055336, -0.05087492614984512, -0.05229071155190468, 0.017874525859951973, 0.030306633561849594, -0.08637596666812897, -0.14898031949996948, -0.033717166632413864, 0.03549951687455177, -0.02099117636680603, -0.09959717094898224, 0.1425362080335617, 0.009977969340980053, 0.030099906027317047, 0.06006331741809845, -0.1419394314289093, -0.11054148524999619, -0.05316055193543434, 0.1764623522758484, -0.013841159641742706, 0.08491477370262146, -0.0679752379655838, -0.027208346873521805, 0.0776199921965599, 0.005507218651473522, -0.07939522713422775, -0.0956752821803093, 0.014004907570779324, -0.033804744482040405, 0.047285307198762894, 0.003760363906621933, 0.10455207526683807, 0.1213613748550415, -0.060092952102422714, -0.09161023795604706, 0.08383911848068237, 0.06145409122109413, -0.03175671771168709, 0.0024884003214538097, 0.036304302513599396, 0.01577497087419033, 0.006489275489002466, 0.0643385723233223, -0.07827051728963852, 0.04166750609874725, -0.05260009691119194, 0.008684852160513401, -0.15147928893566132, 0.08150976896286011, -0.1063992977142334, 0.1065467968583107, 0.08419342339038849, -0.05307234078645706, 0.013027235865592957, -0.053080152720212936, 0.038661323487758636, 0.047524187713861465, 0.0027155212592333555, -0.01451865117996931, -0.15680384635925293, -0.061711084097623825, 0.0014604139141738415, 0.061572663486003876, -0.2381313145160675, -0.04001330956816673, -0.13010317087173462, 0.02616901509463787, -0.037271711975336075, 0.0698726698756218, 0.11954330652952194, 0.01583435758948326, 0.02147742174565792, -0.27362802624702454, 0.015965139493346214, 0.1094847172498703, -0.09396284818649292, -0.04051069915294647 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"license": "apache-2.0", "library_name": "transformers", "datasets": ["jondurbin/truthy-dpo-v0.1"]}
text-generation
vicgalle/CarbonBeagle-11B-truthy
[ "transformers", "safetensors", "mistral", "text-generation", "conversational", "dataset:jondurbin/truthy-dpo-v0.1", "arxiv:1910.09700", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T12:40:49+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #mistral #text-generation #conversational #dataset-jondurbin/truthy-dpo-v0.1 #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #mistral #text-generation #conversational #dataset-jondurbin/truthy-dpo-v0.1 #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 84, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #mistral #text-generation #conversational #dataset-jondurbin/truthy-dpo-v0.1 #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]" ]
[ -0.07881546765565872, 0.19828130304813385, -0.003788086585700512, 0.02594563364982605, 0.10099275410175323, -0.004296048078685999, 0.0479087196290493, 0.1277962476015091, 0.00588206248357892, 0.13631753623485565, 0.035890862345695496, 0.15719355642795563, 0.10844240337610245, 0.1793917864561081, 0.007658287417143583, -0.15868471562862396, 0.0245195385068655, -0.10021348297595978, 0.024458106607198715, 0.11435500532388687, 0.12693296372890472, -0.10097949206829071, 0.06654553860425949, -0.04340309277176857, 0.005332469940185547, -0.05228358507156372, -0.07581142336130142, -0.03051694668829441, 0.04182282090187073, 0.029649484902620316, 0.05053135007619858, -0.004885184578597546, 0.07960651814937592, -0.2893868088722229, 0.019171148538589478, 0.05957712233066559, -0.0068930331617593765, 0.081367626786232, 0.08080241084098816, -0.04768414795398712, 0.052873868495225906, -0.09125132113695145, 0.11165368556976318, 0.1035337820649147, -0.07134167850017548, -0.15685921907424927, -0.0693986713886261, 0.10982432961463928, 0.17701099812984467, 0.061307866126298904, -0.03139302507042885, 0.1258116364479065, -0.029040049761533737, 0.03488624840974808, 0.042504508048295975, -0.06233052536845207, -0.054855868220329285, 0.05151001363992691, 0.12500938773155212, 0.05076473951339722, -0.12248334288597107, 0.0024265197571367025, 0.028647491708397865, 0.027724992483854294, 0.09080187231302261, 0.014314192347228527, 0.15132085978984833, 0.03204616904258728, -0.13304442167282104, -0.04573747143149376, 0.06554054468870163, 0.019732391461730003, -0.03674599528312683, -0.24686864018440247, -0.011515004560351372, -0.03573618829250336, -0.03669242933392525, -0.07523033767938614, 0.04231909289956093, 0.010482934303581715, 0.10053340345621109, -0.03943395987153053, -0.079249806702137, -0.020038429647684097, 0.07347044348716736, 0.058268915861845016, 0.015719229355454445, -0.032701823860406876, 0.007287043146789074, 0.10275758057832718, 0.07646103203296661, -0.13315226137638092, -0.047502677887678146, -0.05886125564575195, -0.0792907103896141, -0.03181007131934166, 0.049969371408224106, 0.058802664279937744, 0.056187473237514496, 0.23812977969646454, 0.00020414328901097178, 0.04227973148226738, 0.017442164942622185, -0.0027322338428348303, 0.05583572015166283, 0.0890578180551529, -0.06800514459609985, -0.16957883536815643, -0.018976964056491852, 0.0959894210100174, -0.0005425198469310999, -0.038048259913921356, -0.05217256024479866, 0.052499543875455856, 0.06655170768499374, 0.11139781773090363, 0.15648657083511353, 0.004293535836040974, -0.07813064008951187, -0.07859116047620773, 0.2078639417886734, -0.14715109765529633, 0.034201689064502716, 0.00617571035400033, -0.015117393806576729, -0.04204301908612251, 0.010914871469140053, 0.025545824319124222, -0.043178070336580276, 0.061957795172929764, -0.06763803958892822, -0.045910272747278214, -0.10450239479541779, -0.014932607300579548, 0.04488568380475044, -0.01481498870998621, -0.03474658355116844, -0.04571165889501572, -0.08340682089328766, -0.084851935505867, 0.09068591892719269, -0.06649108231067657, -0.05018448084592819, -0.03816819190979004, -0.08369343727827072, 0.030897997319698334, 0.004291737452149391, 0.05552113056182861, -0.024042882025241852, 0.0485309474170208, -0.023935697972774506, 0.043150316923856735, 0.09938269108533859, 0.039635591208934784, -0.06291232258081436, 0.06383609771728516, -0.17414671182632446, 0.0871361494064331, -0.07548221200704575, 0.038742970675230026, -0.16294081509113312, -0.008647636510431767, 0.050708867609500885, 0.01890251226723194, 0.016494527459144592, 0.15003183484077454, -0.1978839486837387, -0.011560849845409393, 0.16447751224040985, -0.08672865480184555, -0.1205868273973465, 0.038608577102422714, -0.04632066935300827, 0.17060771584510803, 0.04823600500822067, -0.026249073445796967, 0.08270324766635895, -0.15326376259326935, -0.07186512649059296, -0.04507352411746979, -0.017185712233185768, 0.11828253418207169, 0.07537685334682465, -0.06942427903413773, 0.07968953251838684, 0.027863051742315292, -0.050339024513959885, -0.04515686258673668, -0.027160923928022385, -0.10464459657669067, 0.02292945794761181, -0.0703677237033844, 0.033380892127752304, -0.005213726311922073, -0.07840840518474579, 0.004490640480071306, -0.17269867658615112, -0.014996840618550777, 0.08547213673591614, 0.010901116766035557, -0.012243608944118023, -0.100921630859375, 0.013962926343083382, -0.010242554359138012, 0.0026573913637548685, -0.12942355871200562, -0.05195658653974533, 0.03399115055799484, -0.13868989050388336, 0.015431039966642857, -0.1263042688369751, 0.05113783851265907, 0.01730240322649479, -0.04055362194776535, -0.03885255753993988, 0.012320130132138729, 0.013627488166093826, -0.03197228163480759, -0.23036736249923706, -0.03626461327075958, -0.05209866166114807, 0.12405294179916382, -0.17210176587104797, 0.04660525172948837, 0.03528955951333046, 0.154791459441185, -0.008823263458907604, -0.062012650072574615, 0.023054346442222595, -0.06283076107501984, -0.028296422213315964, -0.0581851452589035, 0.016212422400712967, -0.03035099245607853, -0.036525145173072815, 0.029630037024617195, -0.1738484650850296, -0.07187142968177795, 0.10492879152297974, 0.037480615079402924, -0.11284895241260529, -0.04220625385642052, -0.017289940267801285, -0.0842842236161232, -0.027866095304489136, -0.09179391711950302, 0.11602558940649033, 0.064602330327034, 0.025295423343777657, -0.05807238444685936, -0.0957917794585228, 0.01143091544508934, 0.005810038652271032, -0.024226972833275795, 0.07300575822591782, 0.03304094448685646, -0.1464795023202896, 0.10047832131385803, 0.08730784058570862, 0.07242788374423981, 0.10477128624916077, -0.010574106127023697, -0.08535327762365341, -0.054289668798446655, 0.053848814219236374, 0.028533045202493668, 0.12984049320220947, -0.08748336136341095, 0.026734469458460808, 0.03565197438001633, -0.046436019241809845, 0.04171659052371979, -0.05022921413183212, 0.02891300618648529, 0.002196888206526637, -0.0008687818772159517, 0.06161680445075035, -0.047712672501802444, 0.00358104077167809, 0.053274258971214294, 0.07516757398843765, 0.03713718429207802, 0.03479282930493355, -0.049143511801958084, -0.11213786154985428, 0.14144432544708252, -0.11473142355680466, -0.22945551574230194, -0.15714401006698608, 0.012630180455744267, 0.027595121413469315, -0.012676641345024109, 0.0002522561408113688, -0.0617019459605217, -0.09118322283029556, -0.06718011945486069, 0.028629612177610397, 0.030481429770588875, -0.07362861186265945, -0.05192512273788452, 0.0673287957906723, 0.03888237103819847, -0.13425830006599426, 0.02333889715373516, 0.05814511328935623, -0.0503443107008934, -0.01452688779681921, 0.09098145365715027, 0.10071311891078949, 0.15509368479251862, 0.026995470747351646, -0.024426763877272606, 0.03903128206729889, 0.1890534609556198, -0.13888412714004517, 0.10429771989583969, 0.14118772745132446, -0.06478054076433182, 0.08099202811717987, 0.21397748589515686, 0.028014136478304863, -0.08075498789548874, 0.02717246115207672, 0.03598237410187721, -0.0168061014264822, -0.2405443638563156, -0.07052956521511078, -0.009542427957057953, -0.06399513781070709, 0.07753880321979523, 0.06916216015815735, 0.10138556361198425, 0.020266665145754814, -0.0951237678527832, -0.06340327113866806, 0.04275375232100487, 0.11852041631937027, -0.015316975302994251, -0.0040738522075116634, 0.08654475957155228, 0.000752968539018184, 0.009894811548292637, 0.09171147644519806, 0.005475899670273066, 0.15728604793548584, 0.029750864952802658, 0.18468399345874786, 0.08047284930944443, 0.08130007237195969, -0.024692878127098083, 0.022424111142754555, 0.034270837903022766, 0.04760831594467163, 0.0010454743169248104, -0.09034740924835205, -0.006805452983826399, 0.1309634894132614, 0.02666550688445568, 0.00850677490234375, 0.027163676917552948, -0.028908856213092804, 0.06819561868906021, 0.17373552918434143, -0.020335545763373375, -0.18999046087265015, -0.07895176857709885, 0.0819571241736412, -0.07079801708459854, -0.12973712384700775, -0.015118171460926533, 0.019867567345499992, -0.15224464237689972, 0.012756898067891598, -0.047877758741378784, 0.10251205414533615, -0.11504417657852173, -0.011451746337115765, 0.0647401288151741, 0.0570414662361145, 0.01916520670056343, 0.05474560707807541, -0.149484783411026, 0.0977819636464119, 0.037385232746601105, 0.07960565388202667, -0.09108329564332962, 0.09984303265810013, 0.00465016532689333, -0.05566259101033211, 0.174989253282547, -0.006504591554403305, -0.03550473228096962, -0.09755636751651764, -0.12537525594234467, -0.029732022434473038, 0.11871903389692307, -0.1574956625699997, 0.09242075681686401, -0.030641116201877594, -0.035121552646160126, 0.0009906096383929253, -0.07698814570903778, -0.1236131489276886, -0.1708577573299408, 0.07721659541130066, -0.12350669503211975, 0.048636823892593384, -0.09172387421131134, -0.05109061300754547, -0.003261775244027376, 0.2252921313047409, -0.23203054070472717, -0.08419772982597351, -0.1426885426044464, -0.07339078933000565, 0.15236058831214905, -0.06525767594575882, 0.08257818967103958, -0.0021147867664694786, 0.13715380430221558, 0.001774284290149808, -0.025268927216529846, 0.08819132298231125, -0.089314766228199, -0.17512951791286469, -0.0634731575846672, 0.14017701148986816, 0.12897665798664093, 0.021026894450187683, -0.008259249851107597, 0.028918003663420677, -0.025781404227018356, -0.11042740195989609, 0.024032382294535637, 0.18602684140205383, 0.08706289529800415, -0.003113867249339819, -0.026645395904779434, -0.15675631165504456, -0.07985081523656845, -0.042571332305669785, -0.00569736585021019, 0.1937917321920395, -0.06679373979568481, 0.16908712685108185, 0.1621122658252716, -0.06503459811210632, -0.2172807902097702, -0.0216207355260849, 0.034951429814100266, -0.009944216348230839, 0.03756925091147423, -0.17392128705978394, 0.09013393521308899, -0.04237564280629158, -0.06530312448740005, 0.11577162891626358, -0.1340962052345276, -0.14075429737567902, 0.07945678383111954, 0.03325343504548073, -0.17837609350681305, -0.13257193565368652, -0.11716437339782715, -0.013849650509655476, -0.12169323116540909, 0.08157778531312943, -0.01762000471353531, 0.00006085873246775009, 0.024526460096240044, 0.005087598692625761, 0.039726510643959045, -0.05819311738014221, 0.17777453362941742, -0.023438790813088417, 0.00033240654738619924, -0.07579733431339264, -0.09002464264631271, 0.06345563381910324, -0.05954119563102722, 0.09047460556030273, -0.01610659249126911, 0.021563906222581863, -0.10401412099599838, -0.04758157953619957, -0.05587709695100784, 0.026472527533769608, -0.09847619384527206, -0.10183899849653244, -0.019374975934624672, 0.09902661293745041, 0.10612063854932785, -0.017491566017270088, 0.037717778235673904, -0.058098189532756805, 0.05849234759807587, 0.2573933005332947, 0.18561600148677826, 0.06237258389592171, -0.04560523480176926, -0.004375700373202562, -0.02589179389178753, 0.045906614512205124, -0.1884150356054306, 0.050544463098049164, 0.05442243441939354, 0.019697144627571106, 0.07854799181222916, -0.01410416979342699, -0.154921293258667, -0.061627428978681564, 0.07190555334091187, -0.05120985954999924, -0.18615977466106415, -0.026090208441019058, 0.037609562277793884, -0.1925758421421051, -0.019302330911159515, 0.07155706733465195, 0.009145748801529408, -0.037576209753751755, 0.0196884423494339, 0.09776037186384201, -0.0009301623213104904, 0.08760736137628555, 0.05622025579214096, 0.08715860545635223, -0.08696423470973969, 0.06254485994577408, 0.10142189264297485, -0.05448020249605179, 0.04566512629389763, 0.1121007651090622, -0.049286868423223495, -0.04838484153151512, 0.06194585934281349, 0.05679060146212578, 0.019992122426629066, -0.031451266258955, 0.0070560690946877, -0.011169198900461197, 0.04698837175965309, 0.08127573132514954, 0.01714867353439331, 0.007729835342615843, 0.06085601821541786, 0.045793913304805756, -0.05725101754069328, 0.12884485721588135, 0.04548371955752373, 0.015413968823850155, -0.05959885194897652, -0.036754339933395386, 0.0016637861263006926, -0.012131733819842339, -0.02155674621462822, -0.011944356374442577, -0.08570700883865356, -0.013417158275842667, -0.15673376619815826, 0.030557289719581604, -0.10762830078601837, -0.00041262651211582124, 0.006363848224282265, -0.030188586562871933, 0.009064258076250553, 0.01401964109390974, -0.05151880159974098, -0.08674784004688263, -0.015339035540819168, 0.11590209603309631, -0.16002780199050903, 0.014311563223600388, 0.07994300872087479, -0.09922817349433899, 0.08586767315864563, -0.012157579883933067, 0.008687136694788933, 0.010018114000558853, -0.14188933372497559, 0.05494757741689682, -0.040540438145399094, 0.00593084841966629, 0.008361957967281342, -0.1925685554742813, -0.016736449673771858, -0.03982504829764366, -0.05960344895720482, -0.002206806093454361, 0.004871213808655739, -0.11585628986358643, 0.06224488839507103, 0.011133942753076553, -0.04489300027489662, -0.0331563763320446, 0.03556939959526062, 0.09809030592441559, -0.03379448875784874, 0.09875980764627457, -0.007952249608933926, 0.07526570558547974, -0.17710962891578674, -0.003297174582257867, -0.013335742056369781, 0.04430251941084862, -0.012741725891828537, -0.02896476909518242, 0.04576569050550461, -0.024412399157881737, 0.14052896201610565, -0.04272871091961861, 0.0731034129858017, 0.04841770604252815, 0.00845903530716896, -0.0006744765560142696, 0.08067823201417923, 0.04936865717172623, 0.010209578089416027, 0.010102658532559872, 0.009001818485558033, -0.0203599501401186, -0.05675429105758667, -0.16490603983402252, 0.016566429287195206, 0.1838306337594986, 0.09017551690340042, 0.0029047399293631315, 0.06561041623353958, -0.12029856443405151, -0.11793289333581924, 0.13760054111480713, -0.03872014954686165, -0.03176084905862808, -0.08670827746391296, 0.12873926758766174, 0.1325829029083252, -0.17378957569599152, 0.08253435790538788, -0.05428522825241089, -0.041164152324199677, -0.0871848538517952, -0.20769929885864258, -0.05384355038404465, -0.03117726370692253, -0.014087420888245106, -0.05315567925572395, 0.05790875479578972, 0.0551043339073658, -0.020521970465779305, -0.013969428837299347, 0.08262757956981659, 0.01174824871122837, -0.02660449780523777, 0.04565690830349922, 0.056925710290670395, 0.013070810586214066, -0.0644959956407547, 0.025732891634106636, -0.0058279032818973064, 0.059858012944459915, 0.07507255673408508, 0.02485421858727932, -0.03682160750031471, 0.017871694639325142, -0.028778541833162308, -0.12587258219718933, 0.04013444110751152, -0.012291443534195423, -0.057522859424352646, 0.1866769939661026, 0.027421627193689346, -0.0016996754566207528, -0.01473489310592413, 0.2264416217803955, -0.07107027620077133, -0.08879691362380981, -0.13616283237934113, 0.060508206486701965, -0.04538299888372421, 0.034724973142147064, 0.028873175382614136, -0.11759581416845322, 0.00371868209913373, 0.15012221038341522, 0.1376749873161316, -0.010608833283185959, 0.011741950176656246, 0.042162857949733734, 0.007835287600755692, -0.04732910916209221, 0.024009786546230316, 0.043765999376773834, 0.1749413162469864, -0.06099776551127434, 0.07447808980941772, 0.019976118579506874, -0.10050734132528305, -0.014391222968697548, 0.07854747027158737, -0.03487072139978409, 0.0376862958073616, -0.05841081589460373, 0.11774079501628876, -0.06283203512430191, -0.23269148170948029, 0.038614049553871155, -0.09184236079454422, -0.12224751710891724, -0.036600515246391296, 0.06463715434074402, -0.026931094005703926, 0.016587495803833008, 0.09203383326530457, -0.03750869631767273, 0.1967022716999054, 0.02298274077475071, -0.07267346233129501, -0.03440467268228531, 0.048213981091976166, -0.14733335375785828, 0.2720259726047516, 0.001460630097426474, 0.04151538386940956, 0.118688203394413, -0.017858754843473434, -0.1480560600757599, -0.006574925500899553, 0.09749066084623337, -0.08423356711864471, 0.06818953901529312, 0.2088891863822937, -0.010896257124841213, 0.11990156024694443, 0.07482453435659409, -0.055415112525224686, 0.024019334465265274, -0.05523483827710152, -0.07215474545955658, -0.11175215989351273, 0.08484652638435364, -0.0837579220533371, 0.16375470161437988, 0.10128376632928848, -0.06383172422647476, 0.0015340227400884032, -0.033121850341558456, 0.06503324210643768, -0.008601132780313492, 0.12789203226566315, -0.008612642996013165, -0.19350647926330566, 0.032298408448696136, 0.037629734724760056, 0.11666148900985718, -0.18992139399051666, -0.07883117347955704, 0.07059340178966522, -0.004890486132353544, -0.07717996835708618, 0.11404167860746384, 0.06026830896735191, 0.01665038801729679, -0.03991955891251564, -0.08306323736906052, -0.018458427861332893, 0.12130540609359741, -0.1141250804066658, -0.024323677644133568 ]
null
null
sample-factory
A(n) **APPO** model trained on the **doom_health_gathering_supreme** environment. This model was trained using Sample-Factory 2.0: https://github.com/alex-petrenko/sample-factory. Documentation for how to use Sample-Factory can be found at https://www.samplefactory.dev/ ## Downloading the model After installing Sample-Factory, download the model with: ``` python -m sample_factory.huggingface.load_from_hub -r IrinaArcadievna/rl_course_vizdoom_health_gathering_supreme ``` ## Using the model To run the model after download, use the `enjoy` script corresponding to this environment: ``` python -m .usr.local.lib.python3.10.dist-packages.colab_kernel_launcher --algo=APPO --env=doom_health_gathering_supreme --train_dir=./train_dir --experiment=rl_course_vizdoom_health_gathering_supreme ``` You can also upload models to the Hugging Face Hub using the same script with the `--push_to_hub` flag. See https://www.samplefactory.dev/10-huggingface/huggingface/ for more details ## Training with this model To continue training with this model, use the `train` script corresponding to this environment: ``` python -m .usr.local.lib.python3.10.dist-packages.colab_kernel_launcher --algo=APPO --env=doom_health_gathering_supreme --train_dir=./train_dir --experiment=rl_course_vizdoom_health_gathering_supreme --restart_behavior=resume --train_for_env_steps=10000000000 ``` Note, you may have to adjust `--train_for_env_steps` to a suitably high number as the experiment will resume at the number of steps it concluded at.
{"library_name": "sample-factory", "tags": ["deep-reinforcement-learning", "reinforcement-learning", "sample-factory"], "model-index": [{"name": "APPO", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "doom_health_gathering_supreme", "type": "doom_health_gathering_supreme"}, "metrics": [{"type": "mean_reward", "value": "11.48 +/- 5.52", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
IrinaArcadievna/rl_course_vizdoom_health_gathering_supreme
[ "sample-factory", "tensorboard", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
2024-02-10T12:41:40+00:00
[]
[]
TAGS #sample-factory #tensorboard #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
A(n) APPO model trained on the doom_health_gathering_supreme environment. This model was trained using Sample-Factory 2.0: URL Documentation for how to use Sample-Factory can be found at URL ## Downloading the model After installing Sample-Factory, download the model with: ## Using the model To run the model after download, use the 'enjoy' script corresponding to this environment: You can also upload models to the Hugging Face Hub using the same script with the '--push_to_hub' flag. See URL for more details ## Training with this model To continue training with this model, use the 'train' script corresponding to this environment: Note, you may have to adjust '--train_for_env_steps' to a suitably high number as the experiment will resume at the number of steps it concluded at.
[ "## Downloading the model\n\nAfter installing Sample-Factory, download the model with:", "## Using the model\n\nTo run the model after download, use the 'enjoy' script corresponding to this environment:\n\n\n\nYou can also upload models to the Hugging Face Hub using the same script with the '--push_to_hub' flag.\nSee URL for more details", "## Training with this model\n\nTo continue training with this model, use the 'train' script corresponding to this environment:\n\n\nNote, you may have to adjust '--train_for_env_steps' to a suitably high number as the experiment will resume at the number of steps it concluded at." ]
[ "TAGS\n#sample-factory #tensorboard #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n", "## Downloading the model\n\nAfter installing Sample-Factory, download the model with:", "## Using the model\n\nTo run the model after download, use the 'enjoy' script corresponding to this environment:\n\n\n\nYou can also upload models to the Hugging Face Hub using the same script with the '--push_to_hub' flag.\nSee URL for more details", "## Training with this model\n\nTo continue training with this model, use the 'train' script corresponding to this environment:\n\n\nNote, you may have to adjust '--train_for_env_steps' to a suitably high number as the experiment will resume at the number of steps it concluded at." ]
[ 34, 19, 59, 67 ]
[ "passage: TAGS\n#sample-factory #tensorboard #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n## Downloading the model\n\nAfter installing Sample-Factory, download the model with:## Using the model\n\nTo run the model after download, use the 'enjoy' script corresponding to this environment:\n\n\n\nYou can also upload models to the Hugging Face Hub using the same script with the '--push_to_hub' flag.\nSee URL for more details## Training with this model\n\nTo continue training with this model, use the 'train' script corresponding to this environment:\n\n\nNote, you may have to adjust '--train_for_env_steps' to a suitably high number as the experiment will resume at the number of steps it concluded at." ]
[ -0.162887305021286, -0.07949446886777878, 0.0013769814977422357, 0.0244897473603487, 0.13643795251846313, 0.08826540410518646, 0.13243556022644043, 0.07938782125711441, 0.19449298083782196, 0.07451266050338745, 0.12160012871026993, 0.06742649525403976, 0.02505551464855671, 0.31084391474723816, 0.08655242621898651, -0.18235880136489868, 0.031082456931471825, -0.06436605006456375, -0.02882574498653412, 0.05590416118502617, 0.050910040736198425, -0.06422623991966248, 0.11641133576631546, -0.05714287608861923, -0.15497641265392303, 0.08288847655057907, 0.008126083761453629, 0.03596968948841095, 0.12199652194976807, -0.007729834411293268, 0.06358569860458374, 0.02508161962032318, 0.09885215014219284, -0.08979995548725128, 0.05817115306854248, 0.037268251180648804, -0.005583701189607382, 0.0697544738650322, -0.02916712686419487, 0.01197513286024332, 0.20552261173725128, 0.051445573568344116, -0.014811687171459198, 0.0707944929599762, -0.04854035750031471, 0.005004523321986198, 0.024828260764479637, 0.08118943125009537, 0.1108563020825386, -0.013300174847245216, -0.015604399144649506, 0.2098497599363327, -0.045419543981552124, 0.030687451362609863, 0.1803472340106964, -0.13901305198669434, -0.00587898213416338, 0.3598267436027527, 0.13591337203979492, 0.07389762997627258, -0.05572221428155899, 0.065569669008255, 0.12957775592803955, -0.013377981260418892, -0.022062024101614952, -0.037468962371349335, 0.01014290377497673, 0.02470328100025654, -0.08271043002605438, -0.03898613899946213, 0.18779566884040833, 0.027798498049378395, -0.0647122785449028, -0.11388745903968811, -0.08383605629205704, -0.01143614575266838, -0.08729266375303268, -0.06047317758202553, 0.061255209147930145, 0.06450130045413971, -0.05541218817234039, -0.16354843974113464, -0.08759765326976776, -0.14808951318264008, 0.09711641818284988, -0.018818290904164314, 0.020023507997393608, 0.039053402841091156, -0.13240769505500793, 0.13932685554027557, -0.12239529192447662, -0.005040881223976612, -0.00391974626109004, -0.10012788325548172, -0.0298643596470356, -0.02757178619503975, -0.06954579800367355, -0.08072661608457565, 0.06621979922056198, 0.1397300660610199, 0.1075919046998024, 0.04457515478134155, -0.016096504405140877, 0.0929836705327034, 0.0659836158156395, 0.015487046912312508, -0.046446919441223145, -0.03190334141254425, 0.06750229746103287, 0.09463070333003998, -0.0025161339435726404, -0.04405781999230385, -0.12502750754356384, 0.004669501446187496, -0.05889439582824707, 0.07438734918832779, -0.01944235898554325, 0.09347380697727203, 0.0012449703644961119, -0.0658751055598259, 0.09675891697406769, -0.056166794151067734, -0.015024078078567982, 0.05717969685792923, -0.09829384088516235, -0.044000294059515, 0.02636338584125042, -0.018662840127944946, 0.02191256918013096, -0.08697114139795303, -0.1281215101480484, -0.0406981036067009, -0.15496762096881866, -0.0733695924282074, 0.020342092961072922, -0.10162562131881714, 0.040819648653268814, -0.08701786398887634, -0.27291807532310486, -0.016108427196741104, 0.05915366858243942, 0.0003154690202791244, 0.03663148358464241, -0.06209208071231842, 0.0267410296946764, -0.030988745391368866, -0.013702943921089172, 0.12538094818592072, -0.04706621542572975, 0.005733184050768614, 0.02853262610733509, 0.09092917293310165, 0.029396481812000275, -0.011824010871350765, -0.09237373620271683, 0.03002769686281681, -0.1866937130689621, 0.0038047281559556723, -0.051012441515922546, 0.14028684794902802, -0.07785230129957199, -0.0034444157499819994, -0.07691079378128052, 0.06912831217050552, 0.052552226930856705, 0.21963854134082794, -0.22059281170368195, -0.09743031859397888, 0.1902308464050293, -0.09678838402032852, -0.1949385702610016, 0.06732125580310822, -0.03079940192401409, 0.20069970190525055, 0.02597416751086712, 0.1891578733921051, 0.00020795770979020745, -0.25584760308265686, 0.035303130745887756, 0.07686726003885269, -0.2078019231557846, -0.11653494834899902, 0.00783967413008213, 0.04216665402054787, -0.050144799053668976, 0.023388857021927834, -0.07392873615026474, 0.1217033788561821, -0.023950038477778435, -0.021695949137210846, -0.009935722686350346, -0.06940963864326477, -0.039610356092453, 0.012346661649644375, 0.06086154654622078, -0.02202412113547325, -0.025860905647277832, -0.05173748731613159, 0.16720648109912872, -0.0795547217130661, 0.011736705899238586, -0.11241740733385086, 0.1497063785791397, 0.007124151568859816, 0.025635361671447754, -0.0980280190706253, -0.014672551304101944, 0.044151511043310165, 0.08621654659509659, 0.011970171704888344, 0.1326037049293518, 0.06774137914180756, 0.01454958226531744, 0.042493220418691635, -0.004039871972054243, -0.0012205307139083743, -0.10230473428964615, -0.05593033879995346, -0.11311958730220795, -0.11286478489637375, -0.09429361671209335, 0.08868816494941711, -0.20066434144973755, 0.05826579034328461, -0.15120604634284973, 0.047645486891269684, 0.038803353905677795, -0.07772190868854523, 0.05121537670493126, -0.08661998063325882, -0.021283775568008423, -0.08784573525190353, 0.0805407464504242, -0.014386715367436409, -0.08415807038545609, 0.006313080433756113, -0.09094364196062088, -0.08295580744743347, 0.09175937622785568, 0.013830476440489292, 0.0026490744203329086, -0.1170414388179779, -0.04695970565080643, 0.001149212708696723, 0.03873389959335327, -0.0591595321893692, 0.08649469166994095, 0.06776818633079529, 0.09646541625261307, -0.09070473909378052, 0.03797374665737152, -0.020416714251041412, -0.06236580014228821, -0.045745182782411575, 0.014070805162191391, 0.1767948418855667, -0.022993814200162888, -0.01734299771487713, -0.005982444155961275, -0.048861317336559296, 0.20095843076705933, -0.018403954803943634, -0.11935548484325409, 0.0030399553943425417, -0.01395543571561575, -0.017944620922207832, 0.11660698801279068, -0.13726668059825897, -0.05182260647416115, 0.030854813754558563, -0.06529976427555084, 0.10216285288333893, -0.08242622762918472, -0.0392029769718647, -0.05685178562998772, -0.043409593403339386, 0.046979792416095734, 0.12330524623394012, -0.07290767133235931, -0.009151018224656582, -0.047789376229047775, -0.03510203957557678, -0.025379952043294907, -0.05724980682134628, -0.11478709429502487, 0.1582695096731186, 0.002751561114564538, -0.09990474581718445, -0.17415542900562286, -0.08029486984014511, -0.03834356367588043, 0.05337152257561684, -0.034037429839372635, -0.04430336132645607, -0.01500723510980606, -0.07299388945102692, 0.1465158462524414, 0.063304103910923, -0.0472191721200943, -0.01852818764746189, 0.08560720086097717, 0.04456184431910515, -0.15394946932792664, 0.007078593596816063, -0.08948076516389847, -0.08794131129980087, 0.03091353550553322, -0.08061819523572922, 0.012820594012737274, 0.11341627687215805, 0.03525753691792488, 0.02826494723558426, 0.01035099383443594, 0.23537762463092804, -0.0369284451007843, -0.01093987375497818, 0.19019025564193726, 0.0682438537478447, 0.020443644374608994, 0.055847786366939545, 0.027420951053500175, -0.15370461344718933, 0.10424364358186722, 0.012530675157904625, -0.044538769870996475, -0.10689681768417358, -0.04666181653738022, -0.03360101953148842, 0.09803235530853271, 0.12185155600309372, 0.03158954530954361, 0.025155838578939438, 0.096546471118927, 0.02187134325504303, -0.0098390718922019, -0.11183010786771774, 0.05996714532375336, -0.1770814210176468, -0.043808963149785995, 0.00898060668259859, -0.028755301609635353, 0.00010461114288773388, 0.0659034252166748, 0.026660064235329628, 0.12833580374717712, 0.0295290257781744, 0.06181740015745163, 0.0663255974650383, 0.10200989991426468, 0.01538698747754097, 0.1999037265777588, -0.06215142831206322, -0.1075027585029602, -0.03758005052804947, -0.04118350148200989, -0.11916319280862808, 0.12439136207103729, 0.1381523460149765, -0.030515994876623154, -0.06625506281852722, 0.07200724631547928, 0.014589293859899044, 0.08729344606399536, 0.08250882476568222, -0.29115065932273865, -0.034177567809820175, 0.031450141221284866, 0.01114452164620161, -0.04308335855603218, 0.010566305369138718, 0.10542299598455429, -0.07616783678531647, -0.09982791543006897, -0.03972722589969635, 0.1055394783616066, 0.08046542853116989, 0.03702867403626442, -0.10841067880392075, 0.20128826797008514, -0.01744360849261284, 0.07004447281360626, -0.07662706822156906, 0.1728198230266571, 0.018701205030083656, 0.05943213775753975, -0.07497778534889221, -0.009592941962182522, 0.1228223443031311, 0.03374773636460304, 0.09092900156974792, -0.0056656887754797935, -0.09995020180940628, -0.13336431980133057, -0.1216202825307846, 0.024986369535326958, -0.000090524394181557, -0.08169890940189362, 0.03341596573591232, -0.016717763617634773, 0.017487963661551476, -0.0027857583481818438, 0.23440547287464142, -0.18267135322093964, 0.012482558377087116, -0.054521817713975906, 0.02707577496767044, -0.04300008341670036, -0.0709642544388771, -0.027162717655301094, 0.060507629066705704, 0.09744840115308762, 0.07921962440013885, 0.030401866883039474, -0.07419665157794952, 0.1431404948234558, 0.06514685600996017, -0.058246973901987076, -0.01524845976382494, 0.01951364241540432, 0.1256532073020935, -0.07438289374113083, -0.10393836349248886, 0.10585980117321014, -0.11736445128917694, 0.008749126456677914, -0.05019083246588707, 0.04299405962228775, 0.02305823378264904, 0.011290842667222023, 0.007447924464941025, -0.04279239848256111, 0.0015383695717900991, -0.06904047727584839, 0.0778660774230957, 0.020559091120958328, -0.0047941361553967, -0.0006717707728967071, -0.16239388287067413, 0.08390985429286957, -0.04138755425810814, 0.052877847105264664, 0.1489589661359787, 0.27864590287208557, -0.02386910282075405, 0.030926240608096123, 0.1617380678653717, -0.01897917501628399, -0.2491649091243744, 0.04654841497540474, 0.014908025041222572, 0.10310175269842148, 0.04640066251158714, -0.19236695766448975, 0.11111847311258316, 0.009474517777562141, -0.02225719392299652, 0.009804603643715382, -0.24880149960517883, -0.13740544021129608, 0.17525193095207214, 0.06902051717042923, 0.15983323752880096, -0.03665107116103172, -0.013587141409516335, -0.061109546571969986, -0.03419603407382965, -0.026354335248470306, -0.12708203494548798, 0.12749767303466797, -0.017607107758522034, 0.047745801508426666, 0.027817612513899803, -0.07676684111356735, 0.12058744579553604, -0.017944786697626114, 0.13344953954219818, -0.017018258571624756, -0.031023232266306877, 0.042466819286346436, -0.09033756703138351, 0.1662607043981552, -0.10233280807733536, 0.057950668036937714, -0.11091876775026321, -0.03109682910144329, -0.015322481282055378, 0.15654151141643524, 0.005544521380215883, -0.0855189636349678, -0.041066281497478485, 0.04975702613592148, -0.05784251168370247, 0.05022609233856201, -0.0021613158751279116, -0.03506873920559883, 0.022246064618229866, 0.08415499329566956, 0.040208954364061356, -0.10403558611869812, -0.011038471013307571, 0.03089289739727974, 0.01896476000547409, 0.09993185102939606, -0.20835483074188232, -0.020152123644948006, 0.019231827929615974, -0.015702085569500923, 0.13085414469242096, 0.04400704801082611, -0.08080117404460907, 0.027568496763706207, 0.13726983964443207, -0.061186157166957855, -0.030986590310931206, -0.04847807064652443, -0.016679393127560616, -0.12794725596904755, -0.01594163477420807, 0.057148490101099014, -0.04251079633831978, 0.02512725070118904, -0.03424951806664467, 0.0004248716577421874, -0.10717252641916275, 0.07036283612251282, 0.06859682500362396, 0.0642281174659729, -0.07167360186576843, 0.09394960850477219, -0.07811970263719559, 0.014289900660514832, 0.03734226152300835, 0.045441556721925735, -0.06931920349597931, -0.06820165365934372, -0.05322124809026718, 0.27575042843818665, -0.024388493970036507, -0.02025510184466839, -0.06021025776863098, 0.11942195147275925, -0.057836465537548065, -0.06673881411552429, 0.08716115355491638, -0.007450808770954609, -0.059019722044467926, 0.022327717393636703, -0.0734894648194313, -0.014457973651587963, 0.04693116992712021, 0.016375891864299774, -0.11610891669988632, 0.1136312261223793, 0.031648989766836166, 0.02891513518989086, -0.09186926484107971, -0.0486464723944664, -0.12123195827007294, 0.0032020595390349627, -0.025323880836367607, -0.06051601842045784, -0.07913094758987427, -0.0425749197602272, 0.049642790108919144, 0.018434861674904823, -0.08444267511367798, -0.0022111251018941402, -0.12617166340351105, 0.006370943505316973, 0.006689207162708044, 0.10316617041826248, -0.06351965665817261, 0.04670397937297821, 0.10049878805875778, -0.07692139595746994, 0.09893755614757538, 0.0846271738409996, -0.00729260453954339, 0.08929292112588882, -0.20261284708976746, -0.02319980226457119, 0.047821637243032455, 0.055264540016651154, 0.03154374286532402, 0.06104309484362602, 0.013487739488482475, -0.05460033565759659, 0.04538526386022568, -0.03539090231060982, 0.0028435050044208765, -0.09104080498218536, 0.09713591635227203, 0.009731475263834, -0.009716489352285862, -0.060456521809101105, -0.01384128537029028, 0.01817488856613636, 0.10404353588819504, 0.09692291915416718, -0.07237115502357483, -0.0035003575030714273, -0.11786255985498428, 0.024597108364105225, 0.02565017342567444, 0.010576808825135231, 0.03638135641813278, -0.11692339926958084, 0.03729743883013725, -0.05475534871220589, 0.19700418412685394, 0.019796879962086678, -0.10531783103942871, -0.008661900646984577, 0.07250577956438065, 0.17378750443458557, -0.006129021290689707, 0.21011123061180115, 0.05919691175222397, 0.09556611627340317, 0.0324610099196434, 0.11373614519834518, 0.11542147397994995, 0.004254546947777271, 0.10733281821012497, 0.0500684529542923, -0.04822303727269173, 0.14306919276714325, 0.032827045768499374, -0.017670227214694023, 0.0304852481931448, 0.04704435542225838, -0.03187015652656555, 0.02075354754924774, -0.06440161913633347, 0.11196915805339813, 0.13514995574951172, -0.08471442013978958, -0.0081911850720644, 0.04797748476266861, -0.0438203290104866, -0.1532401293516159, -0.08671712130308151, -0.024648865684866905, -0.2236001342535019, 0.08533021807670593, -0.06946314871311188, -0.13578248023986816, 0.019155733287334442, 0.013867083936929703, -0.028145823627710342, 0.11776147037744522, -0.07801362872123718, -0.03346126526594162, 0.020983682945370674, -0.039618294686079025, -0.09754771739244461, -0.09402462840080261, -0.07874704152345657, 0.03500581532716751, -0.04535633698105812, 0.025271590799093246, -0.05421067774295807, 0.015182215720415115, 0.10334893316030502, -0.04038224741816521, -0.041323766112327576, -0.0359976626932621, -0.035855069756507874, -0.11793428659439087, 0.025968458503484726, 0.044103916734457016, -0.03597194701433182, -0.05585090070962906, 0.17637495696544647, -0.04257858544588089, -0.01666315644979477, -0.1211012676358223, 0.14332374930381775, -0.04330325871706009, 0.03261799365282059, -0.10366860777139664, -0.08559805154800415, -0.10071583092212677, 0.27439257502555847, 0.2784624397754669, -0.14349330961704254, -0.009759977459907532, 0.02939503826200962, 0.004204166121780872, -0.14250165224075317, 0.14376720786094666, 0.01570971868932247, -0.024460898712277412, -0.027595078572630882, 0.026391539722681046, -0.007621914613991976, -0.0827714279294014, -0.03114704228937626, -0.05752136558294296, -0.006779014132916927, -0.05148708075284958, -0.034257955849170685, 0.06298708915710449, -0.12136059254407883, -0.09091135859489441, -0.05560125410556793, -0.0083417734131217, -0.03344108536839485, -0.07473809272050858, -0.019548200070858, 0.07662302255630493, 0.14781777560710907, -0.05502733215689659, 0.06005467101931572, -0.004367031157016754, -0.04969286173582077, -0.13970479369163513, -0.13660922646522522, 0.05449144169688225, -0.129489928483963, 0.26909253001213074, -0.050524767488241196, -0.05207161232829094, 0.041712693870067596, -0.03221052139997482, -0.05838879942893982, 0.020522039383649826, 0.009778409264981747, -0.05078497156500816, -0.029240628704428673, 0.09255361557006836, -0.033305004239082336, 0.009149706922471523, -0.022496739402413368, -0.22135144472122192, 0.0034119023475795984, -0.05107501149177551, 0.028507398441433907, -0.12569822371006012, 0.06501629203557968, -0.09348012506961823, 0.12403472512960434, 0.07595156878232956, -0.01166640967130661, -0.036088403314352036, -0.04733064025640488, 0.1257045865058899, 0.08392459154129028, -0.02910126931965351, -0.0870935395359993, -0.16758979856967926, -0.004611360374838114, -0.0011314527364447713, -0.08687946200370789, -0.23090760409832, -0.008421163074672222, -0.031696807593107224, 0.0109195401892066, -0.00838692206889391, 0.12826944887638092, 0.14749252796173096, 0.05249129980802536, 0.016358694061636925, -0.12719306349754333, 0.041898638010025024, 0.08496948331594467, -0.15762199461460114, -0.1707899123430252 ]
null
null
transformers
<div align="center"> <h1> MiniCPM </h1> </div> <p align="center"> <a href="https://shengdinghu.notion.site/MiniCPM-c805a17c5c8046398914e47f0542095a?pvs=4" target="_blank">MiniCPM 技术报告</a><a href="https://shengdinghu.notion.site/MiniCPM-Unveiling-the-Potential-of-End-side-Large-Language-Models-d4d3a8c426424654a4e80e42a711cb20?pvs=4" target="_blank"> Technical Report</a> | <a href="https://github.com/OpenBMB/OmniLMM/" target="_blank">OmniLMM 多模态模型 Multi-modal Model</a> | <a href="https://luca.cn/" target="_blank">CPM-C 千亿模型试用 ~100B Model Trial </a> </p> MiniCPM 是面壁与清华大学自然语言处理实验室共同开源的系列端侧语言大模型,主体语言模型 MiniCPM-2B 仅有 24亿(2.4B)的非词嵌入参数量。 - 经过 SFT 后,MiniCPM 在公开综合性评测集上,MiniCPM 与 Mistral-7B相近(中文、数学、代码能力更优),整体性能超越 Llama2-13B、MPT-30B、Falcon-40B 等模型。 - 经过 DPO 后,MiniCPM 在当前最接近用户体感的评测集 MTBench上,MiniCPM-2B 也超越了 Llama2-70B-Chat、Vicuna-33B、Mistral-7B-Instruct-v0.1、Zephyr-7B-alpha 等众多代表性开源大模型。 - 以 MiniCPM-2B 为基础构建端侧多模态大模型 MiniCPM-V,整体性能在同规模模型中实现最佳,超越基于 Phi-2 构建的现有多模态大模型,在部分评测集上达到与 9.6B Qwen-VL-Chat 相当甚至更好的性能。 - 经过 Int4 量化后,MiniCPM 可在手机上进行部署推理,流式输出速度略高于人类说话速度。MiniCPM-V 也首次跑通了多模态大模型在手机上的部署。 - 一张1080/2080可高效参数微调,一张3090/4090可全参数微调,一台机器可持续训练 MiniCPM,二次开发成本较低。 我们将完全开源MiniCPM-2B的模型参数供学术研究和有限商用,以及训练过程中的所有Checkpoint和大部分非专有数据供模型机理研究。 - 基于MiniCPM-2B的指令微调与人类偏好对**MiniCPM-2B-SFT/DPO。** - 基于MiniCPM-2B的多模态模型**MiniCPM-V**,能力超越基于Phi-2的同参数级别多模态模型**。** - MiniCPM-2B-SFT/DPO的Int4量化版**MiniCPM-2B-SFT/DPO-Int4。** - 基于MLC-LLM、LLMFarm开发的MiniCPM手机端程序,**文本及多模态模型均可在手机端进行推理。** MiniCPM is an End-Size LLM developed by ModelBest Inc. and TsinghuaNLP, with only 2.4B parameters excluding embeddings. - MiniCPM has very close performance compared with Mistral-7B on open-sourced general benchmarks with better ability on Chinese, Mathmetics and Coding after SFT. The overall performance exceeds Llama2-13B, MPT-30B, Falcon-40B, etc. - After DPO, MiniCPM outperforms Llama2-70B-Chat, Vicuna-33B, Mistral-7B-Instruct-v0.1, Zephyr-7B-alpha, etc. on MTBench. - MiniCPM-V, based on MiniCPM-2B, achieves the best overall performance among multimodel models of the same scale, surpassing existing multimodal large models built on Phi-2 and achieving performance comparable to or even better than 9.6B Qwen-VL-Chat on some tasks. - MiniCPM can be deployed and infer on smartphones, and the speed of streaming output is relatively higher than the verbal speed of human. MiniCPM-V is the first multi-modal models that can be deployed on smartphones. - The cost of developing based on MiniCPM is low. Parameter efficient finetuning can be conducted with a single 1080/2080 GPU and full parameter finetuning can be conducted with a 3090/4090 GPU. We release all model parameters for research and limited commercial use. We also release all the checkpoint during training and most public training data for research on model mechanism. - SFT and DPO version based on MiniCPM-2B and human preference: **MiniCPM-2B-SFT/DPO** - The multi-modal model **MiniCPM-V** based on MiniCPM-2B, which outperforms models with similar size, i.e., Phi-2 - The INT4 quantized version **MiniCPM-2B-SFT/DPO-Int4** based on MiniCPM-2B-SFT/DPO - Mobile phone application based on MLC-LLM and LLMFarm. Both language model and multimodel model can conduct inference on smartphones. ### 评测结果 Evaluation Results 详细的评测结果位于[github仓库](https://github.com/OpenBMB/MiniCPM?tab=readme-ov-file#%E8%AF%84%E6%B5%8B%E7%BB%93%E6%9E%9C) Detailed evaluation results are in [github repo](https://github.com/OpenBMB/MiniCPM/blob/main/README-en.md#evaluation-results) 注意:我们发现使用Huggingface生成质量略差于vLLM,因此推荐使用vLLM进行测试。我们正在排查原因。 Notice: We discovered that the quality of Huggingface generation is slightly lower than vLLM, thus benchmarking using vLLM is recommended. We are investigating the cause now. ### 局限性 Limitations - 受限于模型规模,模型可能出现幻觉性问题。其中由于DPO模型生成的回复内容更长,更容易出现幻觉。我们也将持续进行MiniCPM模型的迭代改进; - 为了保证在学术研究用途上模型的通用性,我们未对模型进行任何身份认同训练。同时由于我们用ShareGPT开源语料作为部分训练数据,模型可能会输出类似GPT系列模型的身份认同信息; - 受限于模型规模,模型的输出受到提示词(prompt)的影响较大,可能多次尝试产生不一致的结果; - 受限于模型容量,模型的知识记忆较不准确,后续我们将结合RAG方法来增强模型的知识记忆能力。 - Due to limitations in model size, the model may experience hallucinatory issues. As DPO model tend to generate longer response, hallucinations are more likely to occur. We will also continue to iterate and improve the MiniCPM model. - To ensure the universality of the model for academic research purposes, we did not conduct any identity training on the model. Meanwhile, as we use ShareGPT open-source corpus as part of the training data, the model may output identity information similar to the GPT series models. - Due to the limitation of model size, the output of the model is greatly influenced by prompt words, which may result in inconsistent results from multiple attempts. - Due to limited model capacity, the model's knowledge memory is not accurate. In the future, we will combine the RAG method to enhance the model's knowledge memory ability. ## 模型下载 Download | HuggingFace | ModelScope | WiseModel | |-------------|------------|-----------| |[sft-bf16](https://huggingface.co/openbmb/MiniCPM-2B-sft-bf16)|[sft-bf16](https://modelscope.cn/models/OpenBMB/miniCPM-bf16)|[sft-bf16](https://wisemodel.cn/models/OpenBMB/miniCPM-bf16) |[sft-fp32](https://huggingface.co/openbmb/MiniCPM-2B-sft-fp32)|[sft-fp32](https://modelscope.cn/models/OpenBMB/MiniCPM-2B-sft-fp32)|[sft-fp32](https://wisemodel.cn/models/OpenBMB/miniCPM-dpo-fp32) |[dpo-bf16](https://huggingface.co/openbmb/MiniCPM-2B-dpo-bf16)|[dpo-bf16](https://modelscope.cn/models/OpenBMB/MiniCPM-2B-dpo-bf16/summary)|[dpo-bf16](https://wisemodel.cn/models/OpenBMB/MiniCPM-2B-dpo-bf16) |[dpo-fp16](https://huggingface.co/openbmb/MiniCPM-2B-dpo-fp16)|[dpo-fp16](https://modelscope.cn/models/OpenBMB/MiniCPM-2B-dpo-fp16/)|[dpo-fp16](https://wisemodel.cn/models/OpenBMB/MiniCPM-2B-dpo-fp16) |[dpo-fp32](https://huggingface.co/openbmb/MiniCPM-2B-dpo-fp32)|[dpo-fp32](https://modelscope.cn/models/OpenBMB/MiniCPM-2B-dpo-fp32)|[dpo-fp32](https://wisemodel.cn/models/OpenBMB/miniCPM-dpo-fp32) ## 模型使用 Usage * 安装`transformers>=4.36.0`以及`accelerate`后,运行以下代码 * 注意:需要在`from_pretrained`中明确指明模型的数据类型,否则会引起较大计算误差 * Run the following code after install `transformers>=4.36.0` and `accelerate` * Warning: It is necessary to specify the data type of the model clearly in 'from_pretrained', otherwise large calculation errors will be caused ```python from transformers import AutoModelForCausalLM, AutoTokenizer import torch torch.manual_seed(0) path = 'openbmb/MiniCPM-2B-sft-bf16' tokenizer = AutoTokenizer.from_pretrained(path) model = AutoModelForCausalLM.from_pretrained(path, torch_dtype=torch.bfloat16, device_map='cuda', trust_remote_code=True) responds, history = model.chat(tokenizer, "山东省最高的山是哪座山, 它比黄山高还是矮?差距多少?", temperature=0.8, top_p=0.8) print(responds) ``` * 期望输出 Expected Output ```shell 山东省最高的山是泰山,海拔1545米。 相对于黄山(海拔1864米),泰山海拔较低,相差约319米。 ``` ## 开源协议 LICENSE #### 模型协议 Model LICENSE * 本仓库中代码依照 [Apache-2.0](https://github.com/OpenBMB/MiniCPM/blob/main/LICENSE) 协议开源 * MiniCPM 模型权重的使用则需要遵循 [“通用模型许可协议-来源说明-宣传限制-商业授权”](https://github.com/OpenBMB/General-Model-License/blob/main/%E9%80%9A%E7%94%A8%E6%A8%A1%E5%9E%8B%E8%AE%B8%E5%8F%AF%E5%8D%8F%E8%AE%AE-%E6%9D%A5%E6%BA%90%E8%AF%B4%E6%98%8E-%E5%AE%A3%E4%BC%A0%E9%99%90%E5%88%B6-%E5%95%86%E4%B8%9A%E6%8E%88%E6%9D%83.md)。 * MiniCPM 模型权重对学术研究完全开放。 * 如需将模型用于商业用途,请联系[email protected]来获取书面授权,在登记后亦允许免费商业使用。 * This repository is released under the [Apache-2.0](https://github.com/OpenBMB/MiniCPM/blob/main/LICENSE) License. * The usage of MiniCPM model weights must strictly follow [the General Model License (GML)](https://github.com/OpenBMB/General-Model-License/blob/main/%E9%80%9A%E7%94%A8%E6%A8%A1%E5%9E%8B%E8%AE%B8%E5%8F%AF%E5%8D%8F%E8%AE%AE-%E6%9D%A5%E6%BA%90%E8%AF%B4%E6%98%8E-%E5%AE%A3%E4%BC%A0%E9%99%90%E5%88%B6-%E5%95%86%E4%B8%9A%E6%8E%88%E6%9D%83.md). * The models and weights of MiniCPM are completely free for academic research. * If you intend to utilize the model for commercial purposes, please reach out to [email protected] to obtain the certificate of authorization. #### 声明 Statement * 作为一个语言模型,MiniCPM 通过学习大量的文本来生成内容,但它无法理解、表达个人观点或价值判断,它所输出的任何内容都不代表模型开发者的观点和立场。 * 因此用户在使用 MiniCPM 生成的内容时,应自行负责对其进行评估和验证。 * 如果由于使用 MinCPM 开源模型而导致的任何问题,包括但不限于数据安全问题、公共舆论风险,或模型被误导、滥用、传播或不当利用所带来的任何风险和问题,我们将不承担任何责任。 * As a language model, MiniCPM generates content by learning from a vast amount of text. * However, it does not possess the ability to comprehend or express personal opinions or value judgments. * Any content generated by MiniCPM does not represent the viewpoints or positions of the model developers. * Therefore, when using content generated by MiniCPM, users should take full responsibility for evaluating and verifying it on their own. <p id="8"></p> ## 工作引用 Citation * 如果觉得MiniCPM有助于您的工作,请考虑引用下列[技术报告](https://shengdinghu.notion.site/MiniCPM-c805a17c5c8046398914e47f0542095a?pvs=4) * Please cite our [techinical report](https://shengdinghu.notion.site/MiniCPM-Unveiling-the-Potential-of-End-side-Large-Language-Models-d4d3a8c426424654a4e80e42a711cb20?pvs=4) if you find our work valuable. ``` @inproceedings{minicpm2024, title={MiniCPM:Unveiling the Potential of End-side Large Language Models}, booktitle={OpenBMB Blog}, year={2024} } ```
{"language": ["en", "zh"], "tags": ["MiniCPM", "ModelBest", "THUNLP"]}
null
jncraton/MiniCPM-2B-sft-bf16-llama-format-ct2-int8
[ "transformers", "MiniCPM", "ModelBest", "THUNLP", "en", "zh", "endpoints_compatible", "region:us" ]
2024-02-10T12:44:48+00:00
[]
[ "en", "zh" ]
TAGS #transformers #MiniCPM #ModelBest #THUNLP #en #zh #endpoints_compatible #region-us
MiniCPM ========= [MiniCPM 技术报告](URL target=) [Technical Report](URL target=) | [OmniLMM 多模态模型 Multi-modal Model](URL target=) | [CPM-C 千亿模型试用 ~100B Model Trial](URL target=) MiniCPM 是面壁与清华大学自然语言处理实验室共同开源的系列端侧语言大模型,主体语言模型 MiniCPM-2B 仅有 24亿(2.4B)的非词嵌入参数量。 * 经过 SFT 后,MiniCPM 在公开综合性评测集上,MiniCPM 与 Mistral-7B相近(中文、数学、代码能力更优),整体性能超越 Llama2-13B、MPT-30B、Falcon-40B 等模型。 * 经过 DPO 后,MiniCPM 在当前最接近用户体感的评测集 MTBench上,MiniCPM-2B 也超越了 Llama2-70B-Chat、Vicuna-33B、Mistral-7B-Instruct-v0.1、Zephyr-7B-alpha 等众多代表性开源大模型。 * 以 MiniCPM-2B 为基础构建端侧多模态大模型 MiniCPM-V,整体性能在同规模模型中实现最佳,超越基于 Phi-2 构建的现有多模态大模型,在部分评测集上达到与 9.6B Qwen-VL-Chat 相当甚至更好的性能。 * 经过 Int4 量化后,MiniCPM 可在手机上进行部署推理,流式输出速度略高于人类说话速度。MiniCPM-V 也首次跑通了多模态大模型在手机上的部署。 * 一张1080/2080可高效参数微调,一张3090/4090可全参数微调,一台机器可持续训练 MiniCPM,二次开发成本较低。 我们将完全开源MiniCPM-2B的模型参数供学术研究和有限商用,以及训练过程中的所有Checkpoint和大部分非专有数据供模型机理研究。 * 基于MiniCPM-2B的指令微调与人类偏好对MiniCPM-2B-SFT/DPO。 * 基于MiniCPM-2B的多模态模型MiniCPM-V,能力超越基于Phi-2的同参数级别多模态模型。 * MiniCPM-2B-SFT/DPO的Int4量化版MiniCPM-2B-SFT/DPO-Int4。 * 基于MLC-LLM、LLMFarm开发的MiniCPM手机端程序,文本及多模态模型均可在手机端进行推理。 MiniCPM is an End-Size LLM developed by ModelBest Inc. and TsinghuaNLP, with only 2.4B parameters excluding embeddings. * MiniCPM has very close performance compared with Mistral-7B on open-sourced general benchmarks with better ability on Chinese, Mathmetics and Coding after SFT. The overall performance exceeds Llama2-13B, MPT-30B, Falcon-40B, etc. * After DPO, MiniCPM outperforms Llama2-70B-Chat, Vicuna-33B, Mistral-7B-Instruct-v0.1, Zephyr-7B-alpha, etc. on MTBench. * MiniCPM-V, based on MiniCPM-2B, achieves the best overall performance among multimodel models of the same scale, surpassing existing multimodal large models built on Phi-2 and achieving performance comparable to or even better than 9.6B Qwen-VL-Chat on some tasks. * MiniCPM can be deployed and infer on smartphones, and the speed of streaming output is relatively higher than the verbal speed of human. MiniCPM-V is the first multi-modal models that can be deployed on smartphones. * The cost of developing based on MiniCPM is low. Parameter efficient finetuning can be conducted with a single 1080/2080 GPU and full parameter finetuning can be conducted with a 3090/4090 GPU. We release all model parameters for research and limited commercial use. We also release all the checkpoint during training and most public training data for research on model mechanism. * SFT and DPO version based on MiniCPM-2B and human preference: MiniCPM-2B-SFT/DPO * The multi-modal model MiniCPM-V based on MiniCPM-2B, which outperforms models with similar size, i.e., Phi-2 * The INT4 quantized version MiniCPM-2B-SFT/DPO-Int4 based on MiniCPM-2B-SFT/DPO * Mobile phone application based on MLC-LLM and LLMFarm. Both language model and multimodel model can conduct inference on smartphones. ### 评测结果 Evaluation Results 详细的评测结果位于github仓库 Detailed evaluation results are in github repo 注意:我们发现使用Huggingface生成质量略差于vLLM,因此推荐使用vLLM进行测试。我们正在排查原因。 Notice: We discovered that the quality of Huggingface generation is slightly lower than vLLM, thus benchmarking using vLLM is recommended. We are investigating the cause now. ### 局限性 Limitations * 受限于模型规模,模型可能出现幻觉性问题。其中由于DPO模型生成的回复内容更长,更容易出现幻觉。我们也将持续进行MiniCPM模型的迭代改进; * 为了保证在学术研究用途上模型的通用性,我们未对模型进行任何身份认同训练。同时由于我们用ShareGPT开源语料作为部分训练数据,模型可能会输出类似GPT系列模型的身份认同信息; * 受限于模型规模,模型的输出受到提示词(prompt)的影响较大,可能多次尝试产生不一致的结果; * 受限于模型容量,模型的知识记忆较不准确,后续我们将结合RAG方法来增强模型的知识记忆能力。 * Due to limitations in model size, the model may experience hallucinatory issues. As DPO model tend to generate longer response, hallucinations are more likely to occur. We will also continue to iterate and improve the MiniCPM model. * To ensure the universality of the model for academic research purposes, we did not conduct any identity training on the model. Meanwhile, as we use ShareGPT open-source corpus as part of the training data, the model may output identity information similar to the GPT series models. * Due to the limitation of model size, the output of the model is greatly influenced by prompt words, which may result in inconsistent results from multiple attempts. * Due to limited model capacity, the model's knowledge memory is not accurate. In the future, we will combine the RAG method to enhance the model's knowledge memory ability. 模型下载 Download ------------- HuggingFace: sft-bf16, ModelScope: sft-bf16, WiseModel: sft-bf16 HuggingFace: sft-fp32, ModelScope: sft-fp32, WiseModel: sft-fp32 HuggingFace: dpo-bf16, ModelScope: dpo-bf16, WiseModel: dpo-bf16 HuggingFace: dpo-fp16, ModelScope: dpo-fp16, WiseModel: dpo-fp16 HuggingFace: dpo-fp32, ModelScope: dpo-fp32, WiseModel: dpo-fp32 模型使用 Usage ---------- * 安装'transformers>=4.36.0'以及'accelerate'后,运行以下代码 * 注意:需要在'from\_pretrained'中明确指明模型的数据类型,否则会引起较大计算误差 * Run the following code after install 'transformers>=4.36.0' and 'accelerate' * Warning: It is necessary to specify the data type of the model clearly in 'from\_pretrained', otherwise large calculation errors will be caused * 期望输出 Expected Output 开源协议 LICENSE ------------ #### 模型协议 Model LICENSE * 本仓库中代码依照 Apache-2.0 协议开源 * MiniCPM 模型权重的使用则需要遵循 “通用模型许可协议-来源说明-宣传限制-商业授权”。 * MiniCPM 模型权重对学术研究完全开放。 * 如需将模型用于商业用途,请联系[email protected]来获取书面授权,在登记后亦允许免费商业使用。 * This repository is released under the Apache-2.0 License. * The usage of MiniCPM model weights must strictly follow the General Model License (GML). * The models and weights of MiniCPM are completely free for academic research. * If you intend to utilize the model for commercial purposes, please reach out to cpm@URL to obtain the certificate of authorization. #### 声明 Statement * 作为一个语言模型,MiniCPM 通过学习大量的文本来生成内容,但它无法理解、表达个人观点或价值判断,它所输出的任何内容都不代表模型开发者的观点和立场。 * 因此用户在使用 MiniCPM 生成的内容时,应自行负责对其进行评估和验证。 * 如果由于使用 MinCPM 开源模型而导致的任何问题,包括但不限于数据安全问题、公共舆论风险,或模型被误导、滥用、传播或不当利用所带来的任何风险和问题,我们将不承担任何责任。 * As a language model, MiniCPM generates content by learning from a vast amount of text. * However, it does not possess the ability to comprehend or express personal opinions or value judgments. * Any content generated by MiniCPM does not represent the viewpoints or positions of the model developers. * Therefore, when using content generated by MiniCPM, users should take full responsibility for evaluating and verifying it on their own. 工作引用 Citation ------------- * 如果觉得MiniCPM有助于您的工作,请考虑引用下列技术报告 * Please cite our techinical report if you find our work valuable.
[ "### 评测结果 Evaluation Results\n\n\n详细的评测结果位于github仓库\n\n\nDetailed evaluation results are in github repo\n\n\n注意:我们发现使用Huggingface生成质量略差于vLLM,因此推荐使用vLLM进行测试。我们正在排查原因。\n\n\nNotice: We discovered that the quality of Huggingface generation is slightly lower than vLLM, thus benchmarking using vLLM is recommended.\nWe are investigating the cause now.", "### 局限性 Limitations\n\n\n* 受限于模型规模,模型可能出现幻觉性问题。其中由于DPO模型生成的回复内容更长,更容易出现幻觉。我们也将持续进行MiniCPM模型的迭代改进;\n* 为了保证在学术研究用途上模型的通用性,我们未对模型进行任何身份认同训练。同时由于我们用ShareGPT开源语料作为部分训练数据,模型可能会输出类似GPT系列模型的身份认同信息;\n* 受限于模型规模,模型的输出受到提示词(prompt)的影响较大,可能多次尝试产生不一致的结果;\n* 受限于模型容量,模型的知识记忆较不准确,后续我们将结合RAG方法来增强模型的知识记忆能力。\n* Due to limitations in model size, the model may experience hallucinatory issues. As DPO model tend to generate longer response, hallucinations are more likely to occur. We will also continue to iterate and improve the MiniCPM model.\n* To ensure the universality of the model for academic research purposes, we did not conduct any identity training on the model. Meanwhile, as we use ShareGPT open-source corpus as part of the training data, the model may output identity information similar to the GPT series models.\n* Due to the limitation of model size, the output of the model is greatly influenced by prompt words, which may result in inconsistent results from multiple attempts.\n* Due to limited model capacity, the model's knowledge memory is not accurate. In the future, we will combine the RAG method to enhance the model's knowledge memory ability.\n\n\n模型下载 Download\n-------------\n\n\nHuggingFace: sft-bf16, ModelScope: sft-bf16, WiseModel: sft-bf16\nHuggingFace: sft-fp32, ModelScope: sft-fp32, WiseModel: sft-fp32\nHuggingFace: dpo-bf16, ModelScope: dpo-bf16, WiseModel: dpo-bf16\nHuggingFace: dpo-fp16, ModelScope: dpo-fp16, WiseModel: dpo-fp16\nHuggingFace: dpo-fp32, ModelScope: dpo-fp32, WiseModel: dpo-fp32\n\n\n模型使用 Usage\n----------\n\n\n* 安装'transformers>=4.36.0'以及'accelerate'后,运行以下代码\n* 注意:需要在'from\\_pretrained'中明确指明模型的数据类型,否则会引起较大计算误差\n* Run the following code after install 'transformers>=4.36.0' and 'accelerate'\n* Warning: It is necessary to specify the data type of the model clearly in 'from\\_pretrained', otherwise large calculation errors will be caused\n* 期望输出 Expected Output\n\n\n开源协议 LICENSE\n------------", "#### 模型协议 Model LICENSE\n\n\n* 本仓库中代码依照 Apache-2.0 协议开源\n* MiniCPM 模型权重的使用则需要遵循 “通用模型许可协议-来源说明-宣传限制-商业授权”。\n* MiniCPM 模型权重对学术研究完全开放。\n* 如需将模型用于商业用途,请联系[email protected]来获取书面授权,在登记后亦允许免费商业使用。\n* This repository is released under the Apache-2.0 License.\n* The usage of MiniCPM model weights must strictly follow the General Model License (GML).\n* The models and weights of MiniCPM are completely free for academic research.\n* If you intend to utilize the model for commercial purposes, please reach out to cpm@URL to obtain the certificate of authorization.", "#### 声明 Statement\n\n\n* 作为一个语言模型,MiniCPM 通过学习大量的文本来生成内容,但它无法理解、表达个人观点或价值判断,它所输出的任何内容都不代表模型开发者的观点和立场。\n* 因此用户在使用 MiniCPM 生成的内容时,应自行负责对其进行评估和验证。\n* 如果由于使用 MinCPM 开源模型而导致的任何问题,包括但不限于数据安全问题、公共舆论风险,或模型被误导、滥用、传播或不当利用所带来的任何风险和问题,我们将不承担任何责任。\n* As a language model, MiniCPM generates content by learning from a vast amount of text.\n* However, it does not possess the ability to comprehend or express personal opinions or value judgments.\n* Any content generated by MiniCPM does not represent the viewpoints or positions of the model developers.\n* Therefore, when using content generated by MiniCPM, users should take full responsibility for evaluating and verifying it on their own.\n\n\n\n工作引用 Citation\n-------------\n\n\n* 如果觉得MiniCPM有助于您的工作,请考虑引用下列技术报告\n* Please cite our techinical report if you find our work valuable." ]
[ "TAGS\n#transformers #MiniCPM #ModelBest #THUNLP #en #zh #endpoints_compatible #region-us \n", "### 评测结果 Evaluation Results\n\n\n详细的评测结果位于github仓库\n\n\nDetailed evaluation results are in github repo\n\n\n注意:我们发现使用Huggingface生成质量略差于vLLM,因此推荐使用vLLM进行测试。我们正在排查原因。\n\n\nNotice: We discovered that the quality of Huggingface generation is slightly lower than vLLM, thus benchmarking using vLLM is recommended.\nWe are investigating the cause now.", "### 局限性 Limitations\n\n\n* 受限于模型规模,模型可能出现幻觉性问题。其中由于DPO模型生成的回复内容更长,更容易出现幻觉。我们也将持续进行MiniCPM模型的迭代改进;\n* 为了保证在学术研究用途上模型的通用性,我们未对模型进行任何身份认同训练。同时由于我们用ShareGPT开源语料作为部分训练数据,模型可能会输出类似GPT系列模型的身份认同信息;\n* 受限于模型规模,模型的输出受到提示词(prompt)的影响较大,可能多次尝试产生不一致的结果;\n* 受限于模型容量,模型的知识记忆较不准确,后续我们将结合RAG方法来增强模型的知识记忆能力。\n* Due to limitations in model size, the model may experience hallucinatory issues. As DPO model tend to generate longer response, hallucinations are more likely to occur. We will also continue to iterate and improve the MiniCPM model.\n* To ensure the universality of the model for academic research purposes, we did not conduct any identity training on the model. Meanwhile, as we use ShareGPT open-source corpus as part of the training data, the model may output identity information similar to the GPT series models.\n* Due to the limitation of model size, the output of the model is greatly influenced by prompt words, which may result in inconsistent results from multiple attempts.\n* Due to limited model capacity, the model's knowledge memory is not accurate. In the future, we will combine the RAG method to enhance the model's knowledge memory ability.\n\n\n模型下载 Download\n-------------\n\n\nHuggingFace: sft-bf16, ModelScope: sft-bf16, WiseModel: sft-bf16\nHuggingFace: sft-fp32, ModelScope: sft-fp32, WiseModel: sft-fp32\nHuggingFace: dpo-bf16, ModelScope: dpo-bf16, WiseModel: dpo-bf16\nHuggingFace: dpo-fp16, ModelScope: dpo-fp16, WiseModel: dpo-fp16\nHuggingFace: dpo-fp32, ModelScope: dpo-fp32, WiseModel: dpo-fp32\n\n\n模型使用 Usage\n----------\n\n\n* 安装'transformers>=4.36.0'以及'accelerate'后,运行以下代码\n* 注意:需要在'from\\_pretrained'中明确指明模型的数据类型,否则会引起较大计算误差\n* Run the following code after install 'transformers>=4.36.0' and 'accelerate'\n* Warning: It is necessary to specify the data type of the model clearly in 'from\\_pretrained', otherwise large calculation errors will be caused\n* 期望输出 Expected Output\n\n\n开源协议 LICENSE\n------------", "#### 模型协议 Model LICENSE\n\n\n* 本仓库中代码依照 Apache-2.0 协议开源\n* MiniCPM 模型权重的使用则需要遵循 “通用模型许可协议-来源说明-宣传限制-商业授权”。\n* MiniCPM 模型权重对学术研究完全开放。\n* 如需将模型用于商业用途,请联系[email protected]来获取书面授权,在登记后亦允许免费商业使用。\n* This repository is released under the Apache-2.0 License.\n* The usage of MiniCPM model weights must strictly follow the General Model License (GML).\n* The models and weights of MiniCPM are completely free for academic research.\n* If you intend to utilize the model for commercial purposes, please reach out to cpm@URL to obtain the certificate of authorization.", "#### 声明 Statement\n\n\n* 作为一个语言模型,MiniCPM 通过学习大量的文本来生成内容,但它无法理解、表达个人观点或价值判断,它所输出的任何内容都不代表模型开发者的观点和立场。\n* 因此用户在使用 MiniCPM 生成的内容时,应自行负责对其进行评估和验证。\n* 如果由于使用 MinCPM 开源模型而导致的任何问题,包括但不限于数据安全问题、公共舆论风险,或模型被误导、滥用、传播或不当利用所带来的任何风险和问题,我们将不承担任何责任。\n* As a language model, MiniCPM generates content by learning from a vast amount of text.\n* However, it does not possess the ability to comprehend or express personal opinions or value judgments.\n* Any content generated by MiniCPM does not represent the viewpoints or positions of the model developers.\n* Therefore, when using content generated by MiniCPM, users should take full responsibility for evaluating and verifying it on their own.\n\n\n\n工作引用 Citation\n-------------\n\n\n* 如果觉得MiniCPM有助于您的工作,请考虑引用下列技术报告\n* Please cite our techinical report if you find our work valuable." ]
[ 32, 101, 655, 180, 262 ]
[ "passage: TAGS\n#transformers #MiniCPM #ModelBest #THUNLP #en #zh #endpoints_compatible #region-us \n### 评测结果 Evaluation Results\n\n\n详细的评测结果位于github仓库\n\n\nDetailed evaluation results are in github repo\n\n\n注意:我们发现使用Huggingface生成质量略差于vLLM,因此推荐使用vLLM进行测试。我们正在排查原因。\n\n\nNotice: We discovered that the quality of Huggingface generation is slightly lower than vLLM, thus benchmarking using vLLM is recommended.\nWe are investigating the cause now." ]
[ -0.007029905449599028, -0.18129052221775055, -0.005951940082013607, 0.035357579588890076, 0.09986112266778946, -0.0005699562607333064, 0.046559251844882965, 0.0894036665558815, -0.0011693296255543828, 0.014035903848707676, 0.1396910399198532, 0.05511273443698883, 0.004101620987057686, -0.008708887733519077, 0.04160221666097641, -0.15708091855049133, 0.06503809988498688, 0.045605048537254333, -0.060293711721897125, 0.0694865956902504, 0.167556032538414, -0.0876041129231453, 0.13962098956108093, 0.05953854322433472, -0.15145139396190643, 0.015299891121685505, 0.05207466706633568, 0.040434446185827255, 0.08226493000984192, 0.0536847747862339, -0.020794576033949852, 0.035934820771217346, -0.0060879685916006565, -0.08309119194746017, 0.047387413680553436, 0.07378146797418594, 0.03385411947965622, 0.00824330747127533, -0.0057003977708518505, 0.05097620189189911, 0.18805724382400513, -0.04968774691224098, -0.05933471396565437, 0.06705017387866974, -0.02407228946685791, -0.08207760751247406, -0.0734010636806488, 0.014700877480208874, 0.005885763559490442, 0.01351017877459526, 0.023399267345666885, 0.2902761697769165, -0.14518877863883972, 0.07959070056676865, 0.23669782280921936, -0.17718659341335297, -0.07245728373527527, 0.2521205246448517, 0.0753178521990776, -0.08139427751302719, -0.033760204911231995, 0.13844187557697296, 0.09468132257461548, -0.0026721120811998844, 0.0026779472827911377, 0.0028682169504463673, 0.019566699862480164, 0.036671992391347885, -0.0707022175192833, -0.04395657032728195, 0.21017861366271973, 0.024222983047366142, -0.01676894538104534, -0.07778884470462799, -0.04407808557152748, -0.03170435130596161, 0.019980154931545258, -0.056891828775405884, -0.0009387274621985853, 0.028179902583360672, -0.023554271087050438, -0.056964561343193054, -0.05591902509331703, -0.06277237087488174, -0.050856560468673706, 0.07139258831739426, 0.012281373143196106, 0.06602782756090164, -0.12072503566741943, 0.07584036141633987, -0.016529422253370285, -0.06980713456869125, -0.06994563341140747, -0.15332739055156708, 0.09148465096950531, 0.0343741737306118, -0.029552128165960312, -0.020706990733742714, 0.15407049655914307, 0.06227974221110344, 0.09200252592563629, -0.009465874172747135, 0.01687920093536377, -0.006674732081592083, 0.030380485579371452, 0.16072417795658112, -0.0959244966506958, -0.03150952234864235, 0.11796808242797852, -0.006341414991766214, 0.06780342012643814, -0.010086788795888424, 0.02585897408425808, -0.10685950517654419, 0.0907537117600441, 0.09983225166797638, -0.005054982379078865, 0.08520478010177612, -0.06697309762239456, -0.07043292373418808, -0.13299769163131714, -0.048997700214385986, -0.02080552838742733, -0.02179517038166523, -0.021712658926844597, 0.16008159518241882, 0.037784550338983536, -0.007054767571389675, -0.06459946185350418, -0.04314681142568588, -0.07845790684223175, -0.07143807411193848, -0.07053966075181961, -0.06252117455005646, 0.03135087341070175, -0.12207255512475967, 0.05156022682785988, -0.07779049873352051, -0.11634596437215805, -0.02535473369061947, 0.0038164493162184954, 0.009199460968375206, -0.039813365787267685, 0.056789007037878036, -0.047064024955034256, -0.01596369780600071, 0.005027248989790678, 0.08947260677814484, -0.038541000336408615, 0.02651931159198284, 0.042357850819826126, 0.031899482011795044, -0.08141711354255676, -0.019544366747140884, -0.028555387631058693, 0.04780369997024536, 0.06622783094644547, -0.036718253046274185, -0.10905038565397263, 0.1576538234949112, -0.07052676379680634, -0.06317166239023209, -0.043943632394075394, -0.0029794408474117517, 0.03134134039282799, 0.1293453723192215, -0.13065388798713684, -0.06242143735289574, 0.05081261321902275, -0.12080085277557373, -0.1050158143043518, 0.14760424196720123, 0.009458748623728752, 0.022331831976771355, 0.12191835045814514, -0.0022843179758638144, 0.125326007604599, -0.19225206971168518, -0.020844638347625732, 0.03509785607457161, -0.07853440195322037, 0.0455203503370285, 0.043436478823423386, 0.08068293333053589, -0.14074291288852692, 0.10252724587917328, -0.06006870046257973, 0.007547332439571619, -0.09189192205667496, -0.03843272849917412, -0.15482521057128906, -0.025032080709934235, -0.05366566404700279, 0.04065868258476257, 0.009939560666680336, -0.050259221345186234, -0.04092514142394066, -0.1113775223493576, 0.08663772791624069, -0.023160140961408615, -0.006348270457237959, -0.15282754600048065, 0.1418062448501587, -0.0006764328572899103, 0.03414156287908554, -0.009384505450725555, -0.0311939287930727, 0.01573129929602146, 0.04729769378900528, -0.0020969773177057505, 0.19795012474060059, 0.025106357410550117, -0.06771818548440933, 0.022668274119496346, 0.05191590636968613, -0.025603611022233963, -0.04619232565164566, 0.029351724311709404, -0.1542227417230606, -0.0030155712738633156, -0.04354240745306015, -0.10526778548955917, -0.013441141694784164, -0.02526128478348255, 0.12803754210472107, 0.18103109300136566, 0.03463297337293625, 0.009342518635094166, 0.035754818469285965, -0.04776196554303169, -0.03296038880944252, -0.06930872052907944, 0.05342394486069679, -0.035693034529685974, -0.040911924093961716, 0.0916159525513649, 0.06284727156162262, 0.16049063205718994, 0.12458932399749756, -0.04375477880239487, 0.10487950593233109, 0.03939517214894295, -0.054152533411979675, 0.055077169090509415, -0.022789044305682182, -0.0593547560274601, 0.07723882794380188, 0.0013175575295463204, 0.043476928025484085, -0.0795559361577034, -0.021214991807937622, 0.011689210310578346, -0.05022359639406204, -0.049612510949373245, 0.07128696888685226, 0.06985030323266983, 0.018150771036744118, -0.0025405569467693567, 0.052537064999341965, 0.03218483179807663, 0.21697549521923065, -0.07065971195697784, 0.01958327926695347, -0.018351927399635315, -0.03190876170992851, -0.029256965965032578, 0.06542211771011353, -0.20291361212730408, -0.05130786448717117, 0.04685044288635254, -0.034627415239810944, 0.08495761454105377, -0.09145545959472656, -0.00968084391206503, 0.010166168212890625, -0.01712985895574093, 0.04534405469894409, 0.1705799400806427, -0.018720511347055435, 0.06875555217266083, -0.05606083944439888, 0.10785147547721863, -0.06280707567930222, -0.029891222715377808, -0.05139126256108284, 0.09397685527801514, -0.052178725600242615, -0.1539200097322464, -0.09655670821666718, 0.03096357360482216, -0.15777359902858734, -0.024116067215800285, -0.01893104426562786, -0.20700260996818542, -0.09705216437578201, -0.0003875589754898101, 0.28305497765541077, -0.03400006517767906, 0.06575983017683029, 0.1628442257642746, 0.06021367385983467, 0.014382950030267239, -0.19300132989883423, -0.06491397321224213, -0.01619957946240902, -0.10169330984354019, 0.03864741697907448, -0.06528447568416595, 0.04829725623130798, 0.04979325458407402, 0.030524758622050285, 0.05394037067890167, -0.007211975287646055, 0.32132506370544434, -0.01992679201066494, -0.03821587562561035, 0.36082324385643005, 0.10184833407402039, -0.039524469524621964, 0.05349263548851013, -0.0064017074182629585, -0.14367203414440155, 0.00786982849240303, 0.03457683324813843, -0.07522208243608475, -0.10467099398374557, -0.08188129216432571, -0.08901073783636093, -0.004968418274074793, -0.018996141850948334, 0.03418242558836937, -0.08056186139583588, 0.04029500111937523, -0.03882334381341934, 0.09780184924602509, -0.02714279480278492, 0.03977955877780914, 0.1137450635433197, -0.0636458545923233, 0.09473677724599838, -0.043208952993154526, -0.028461836278438568, 0.11854533851146698, -0.0039004001300781965, 0.1605968028306961, -0.05560247227549553, -0.01632627286016941, 0.1513911783695221, 0.07536110281944275, 0.052956126630306244, 0.04650551453232765, -0.09035848081111908, -0.05243884399533272, -0.016077518463134766, -0.02378683164715767, -0.1326739490032196, -0.03452671691775322, 0.010660284198820591, -0.03295145183801651, -0.07800742983818054, 0.1525101512670517, 0.07630597054958344, 0.1832588016986847, 0.04941178485751152, -0.1264776587486267, -0.055462438613176346, 0.05326181277632713, -0.008437084965407848, -0.051986631006002426, -0.0027289516292512417, 0.021061580628156662, -0.06870332360267639, 0.07071825861930847, -0.03226818889379501, 0.09059399366378784, -0.04277632758021355, 0.08795687556266785, -0.10145540535449982, 0.05973893031477928, 0.029581494629383087, 0.06363088637590408, -0.19256146252155304, 0.11430851370096207, 0.010978278703987598, 0.09448982030153275, -0.08606318384408951, 0.0075591071508824825, 0.07585456222295761, 0.15848422050476074, 0.07580620050430298, 0.018830114975571632, 0.028623787686228752, 0.030678128823637962, -0.046929821372032166, 0.10647550970315933, 0.02444637008011341, 0.018736455589532852, 0.048744652420282364, -0.13624097406864166, 0.06992737203836441, -0.005191147793084383, 0.1014765128493309, -0.19799570739269257, 0.0066639515571296215, 0.02145177498459816, -0.006088297814130783, 0.019627736881375313, -0.10134831815958023, -0.01989709958434105, 0.02259662002325058, 0.1600578874349594, -0.031674597412347794, -0.02505546063184738, -0.10328514128923416, 0.13805516064167023, 0.07889293134212494, -0.03296839818358421, -0.037489306181669235, -0.09881174564361572, 0.14158830046653748, 0.016290584579110146, -0.16540241241455078, -0.08646079897880554, -0.10454490035772324, -0.11585325747728348, 0.027800647541880608, 0.12482665479183197, -0.13428716361522675, 0.012853509746491909, 0.0667235404253006, -0.04668804630637169, -0.07104019820690155, -0.13363845646381378, 0.03306398540735245, 0.18510258197784424, -0.17520250380039215, 0.03133249282836914, -0.01204567402601242, -0.03713098168373108, 0.030830876901745796, -0.031123245134949684, 0.060273922979831696, 0.23087160289287567, -0.014127736911177635, 0.08515125513076782, -0.0011738878674805164, -0.02978982776403427, -0.2783195376396179, -0.042848993092775345, 0.013759949244558811, 0.07922997325658798, -0.09668739885091782, 0.002893011551350355, 0.1082424744963646, -0.057692717760801315, 0.017866728827357292, -0.023136377334594727, -0.23526626825332642, -0.11276296526193619, 0.02768295630812645, 0.04173368960618973, 0.3718765676021576, -0.005618828348815441, -0.018679693341255188, -0.08236056566238403, -0.2331053763628006, 0.09754981845617294, -0.09003683924674988, 0.07901791483163834, -0.09090468287467957, -0.006996639538556337, 0.0546121746301651, -0.04311290755867958, 0.16560278832912445, -0.13848967850208282, 0.09249752014875412, -0.05166751146316528, -0.046906907111406326, 0.07662864029407501, -0.04660817235708237, 0.08431723713874817, -0.00004227057070238516, 0.07783340662717819, -0.10475700348615646, -0.039542991667985916, -0.10952022671699524, 0.11239190399646759, -0.010712068527936935, -0.06655392795801163, -0.14847688376903534, 0.054326292127370834, -0.013305011205375195, 0.007040245458483696, 0.10857025533914566, -0.026810117065906525, 0.006204573903232813, 0.07563278824090958, 0.09349015355110168, -0.25899824500083923, 0.034354791045188904, 0.03968802094459534, 0.04425780847668648, 0.10542070120573044, -0.16066081821918488, 0.07335808128118515, 0.08233226835727692, 0.005068870726972818, -0.00943797454237938, 0.0298149436712265, -0.04710002616047859, 0.006641503889113665, 0.13628563284873962, -0.15587253868579865, -0.04959556460380554, -0.047705113887786865, -0.060472521930933, 0.05329377204179764, 0.046486590057611465, 0.22370930016040802, -0.018752489238977432, 0.011128068901598454, -0.0049996827729046345, 0.003121020970866084, -0.05447318032383919, 0.12865500152111053, 0.03850371018052101, 0.02667682059109211, -0.13138838112354279, 0.04617862030863762, -0.08798189461231232, 0.07230368256568909, -0.012639796361327171, -0.07910933345556259, -0.08561225980520248, -0.04239198938012123, 0.009966548532247543, 0.09304729849100113, -0.03591559827327728, -0.029956577345728874, -0.13225248456001282, -0.1374678760766983, 0.0008156708790920675, 0.003170343115925789, 0.08875831961631775, 0.11562928557395935, -0.054417435079813004, -0.04701881855726242, -0.0401037223637104, 0.004986949265003204, 0.04682082682847977, -0.017912017181515694, -0.21080687642097473, 0.08802185952663422, -0.004020586609840393, 0.10052767395973206, -0.09412624686956406, -0.002272270852699876, -0.1231643483042717, -0.020640898495912552, -0.18347449600696564, -0.04763184115290642, 0.026195168495178223, -0.01202356070280075, 0.0669357180595398, -0.0617840401828289, -0.15457159280776978, 0.0720461755990982, -0.07213762402534485, 0.013440023176372051, -0.020221391692757607, 0.05134095251560211, 0.011415361426770687, 0.029058363288640976, 0.1242644339799881, -0.04318128526210785, 0.07616350054740906, -0.030477555468678474, -0.021776655688881874, 0.045308686792850494, -0.18348194658756256, -0.05518330633640289, 0.08187967538833618, 0.04886586219072342, -0.019504666328430176, 0.023459650576114655, 0.07082602381706238, 0.024681204929947853, 0.00820586085319519, 0.030938245356082916, 0.06416723132133484, -0.05277709290385246, -0.0855041965842247, 0.007049820851534605, 0.008752464316785336, 0.020142750814557076, -0.022155463695526123, 0.04487167298793793, 0.12529303133487701, 0.09726529568433762, 0.014684763737022877, -0.0194696094840765, -0.16838698089122772, 0.00830417312681675, -0.00984482653439045, -0.10293831676244736, -0.028975795954465866, 0.0021882206201553345, 0.06077804043889046, -0.014182863757014275, 0.2449350357055664, 0.11702872812747955, 0.005042670294642448, 0.007772746030241251, 0.18846075236797333, -0.029718095436692238, 0.0003312989429105073, 0.11638941615819931, 0.004579353611916304, 0.13689325749874115, -0.053380146622657776, -0.034658756107091904, 0.06998783349990845, -0.09110356122255325, 0.14839224517345428, 0.04822993278503418, -0.035704005509614944, 0.08797920495271683, 0.05165775865316391, 0.008065287955105305, 0.022829916328191757, 0.014720307663083076, -0.04585376754403114, 0.12946702539920807, -0.03044453077018261, -0.12339377403259277, 0.13701024651527405, -0.09897657483816147, 0.01810203492641449, -0.02869013510644436, -0.006283747963607311, -0.1463201940059662, -0.12284358590841293, -0.07512173801660538, -0.20523324608802795, 0.10039401054382324, -0.10859007388353348, -0.016365915536880493, -0.08122794330120087, 0.05900624021887779, -0.07911669462919235, 0.08075745403766632, -0.08636533468961716, -0.11989813297986984, 0.0785241648554802, -0.024483587592840195, -0.10325786471366882, -0.011768877506256104, 0.03858496621251106, -0.04452567920088768, 0.027312010526657104, -0.02097550593316555, 0.025646530091762543, -0.02509469911456108, -0.08288910984992981, -0.02417762018740177, -0.027694258838891983, -0.06983157992362976, -0.05789756029844284, -0.088569775223732, 0.10396303236484528, -0.028263943269848824, -0.016770722344517708, -0.014409912750124931, 0.16765199601650238, 0.022424353286623955, -0.06278514117002487, -0.12636616826057434, 0.007873471826314926, -0.054353103041648865, 0.03975142538547516, -0.009976043365895748, 0.028603317216038704, -0.04908722639083862, 0.29296132922172546, 0.11499285697937012, -0.06207657605409622, 0.00033675841405056417, -0.07435646653175354, 0.020152198150753975, -0.06945610791444778, 0.17368726432323456, 0.020711664110422134, 0.12320763617753983, -0.03448185324668884, 0.04327443242073059, -0.04210294038057327, -0.029141951352357864, -0.062144361436367035, 0.07125066965818405, 0.07137252390384674, -0.046930860728025436, -0.027866339311003685, 0.03252615034580231, -0.046302780508995056, -0.0053444840013980865, 0.06731386482715607, -0.15821880102157593, -0.046171799302101135, -0.03454064205288887, 0.049210429191589355, -0.06873206794261932, 0.09366195648908615, -0.08513263612985611, 0.03372931480407715, 0.02272685244679451, -0.019690684974193573, -0.1433858722448349, -0.0600271075963974, 0.12623290717601776, 0.1163797527551651, 0.1186923161149025, -0.032800983637571335, -0.03777356073260307, 0.07561182230710983, -0.10685569792985916, -0.08995477855205536, -0.019915208220481873, -0.03216233849525452, -0.12198159843683243, -0.039367012679576874, 0.07561811804771423, 0.02740832418203354, 0.0023442672099918127, 0.02984943799674511, 0.011059924028813839, -0.014475255273282528, 0.12300413101911545, 0.040855977684259415, -0.03510741516947746, 0.13164114952087402, -0.14435340464115143, 0.08194208890199661, 0.1004456952214241, -0.04326357692480087, -0.026686301454901695, -0.11590644717216492, 0.17826007306575775, -0.006448836997151375, 0.01886652782559395, 0.00009278854849981144, -0.1277666836977005, -0.003054513595998287, 0.15814904868602753, -0.07905298471450806, -0.019775869324803352, -0.12718722224235535, -0.02167162485420704, -0.001297742361202836, -0.026628093793988228, 0.13205231726169586, 0.07088743150234222, 0.03847820311784744, 0.053234636783599854, -0.15350842475891113, 0.053478408604860306, -0.016429493203759193, -0.06174127385020256, -0.08909101039171219 ]
null
null
null
# **Q-Learning** Agent playing1 **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="0fluff0/taxi", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
{"tags": ["Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "taxi", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "Taxi-v3", "type": "Taxi-v3"}, "metrics": [{"type": "mean_reward", "value": "7.52 +/- 2.71", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
0fluff0/taxi
[ "Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
2024-02-10T12:45:45+00:00
[]
[]
TAGS #Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
# Q-Learning Agent playing1 Taxi-v3 This is a trained model of a Q-Learning agent playing Taxi-v3 . ## Usage
[ "# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage" ]
[ "TAGS\n#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n", "# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage" ]
[ 32, 33 ]
[ "passage: TAGS\n#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage" ]
[ 0.048862796276807785, -0.16549694538116455, -0.005485367961227894, 0.02960980497300625, 0.1345081776380539, -0.01784728653728962, 0.11895976960659027, 0.07759871333837509, -0.07461097836494446, -0.055395450443029404, 0.1418241262435913, 0.09088201075792313, 0.055222880095243454, 0.05699880048632622, 0.09511256217956543, -0.27440664172172546, 0.048217080533504486, -0.02918700873851776, 0.05621987581253052, 0.11878681182861328, 0.0670095682144165, -0.040441032499074936, 0.061956584453582764, 0.11818158626556396, -0.1018151044845581, -0.007344264071434736, 0.035402704030275345, -0.09440053254365921, 0.17413531243801117, 0.07204403728246689, 0.12337774783372879, 0.05132639780640602, 0.179361954331398, -0.12762396037578583, 0.024310702458024025, -0.0010275895474478602, -0.10138072073459625, -0.03909514099359512, -0.012415820732712746, -0.08349097520112991, 0.03230205550789833, 0.23522862792015076, 0.07199250161647797, 0.06632792949676514, -0.17707863450050354, -0.06584878265857697, -0.04375573247671127, 0.069611094892025, 0.14951466023921967, 0.03758616745471954, -0.033800311386585236, 0.1684885323047638, -0.2564343810081482, 0.05066783353686333, 0.037275806069374084, -0.42313119769096375, 0.017119819298386574, 0.1507398933172226, 0.15090937912464142, 0.06909667700529099, -0.10573802888393402, 0.013512322679162025, 0.051325585693120956, -0.0005318621988408267, 0.024325110018253326, 0.006554204970598221, 0.15601307153701782, 0.08537693321704865, -0.1487821787595749, -0.058576688170433044, 0.17441977560520172, -0.03788546845316887, -0.02613203600049019, -0.039745692163705826, 0.0067160045728087425, -0.06427708268165588, -0.004067842848598957, -0.1777995079755783, 0.00734262028709054, 0.06666424125432968, -0.014348524622619152, 0.014901017770171165, -0.035522811114788055, -0.0966939702630043, -0.023098144680261612, -0.08592145889997482, 0.01677769608795643, -0.006319406442344189, -0.10187895596027374, 0.05002119392156601, -0.061138734221458435, 0.0014382408699020743, -0.05123179033398628, -0.15047866106033325, -0.049055423587560654, -0.03481535613536835, 0.1474713832139969, -0.0044205985032022, -0.01873963139951229, -0.03164304047822952, 0.15474793314933777, 0.049551334232091904, -0.05370146036148071, 0.05625450983643532, 0.07605006545782089, 0.23867930471897125, 0.10401605814695358, 0.10196955502033234, -0.06798075139522552, 0.10180158913135529, -0.12330973148345947, -0.08915644884109497, -0.17508824169635773, 0.11820860952138901, 0.00015364694991149008, 0.1317785084247589, -0.12023144960403442, 0.07898581773042679, -0.067511186003685, 0.013453764840960503, 0.01636839471757412, 0.0820009782910347, -0.012399360537528992, 0.10676060616970062, -0.005061192903667688, -0.06941985338926315, 0.014177112840116024, 0.05935845896601677, 0.03754841163754463, -0.038601722568273544, -0.03192409873008728, -0.05762290954589844, -0.05065649375319481, -0.10128600150346756, -0.06447898596525192, 0.018573462963104248, -0.007677143905311823, -0.1833900660276413, -0.06407523155212402, 0.00897200871258974, 0.015712225809693336, -0.03988850116729736, -0.05148044601082802, -0.15265507996082306, -0.042461175471544266, -0.015450406819581985, -0.03500641882419586, -0.06214277446269989, -0.0383245050907135, 0.046435944736003876, -0.07560601085424423, 0.013364278711378574, 0.023342855274677277, 0.05405820533633232, -0.025881100445985794, 0.06068144738674164, -0.08357544988393784, 0.09493788331747055, -0.1540430635213852, -0.03271956741809845, -0.025445878505706787, -0.041183918714523315, 0.1752462536096573, 0.06099751964211464, -0.015994304791092873, 0.15260063111782074, -0.17141541838645935, -0.058121129870414734, 0.15596486628055573, 0.008629098534584045, -0.09967197477817535, -0.003560945624485612, -0.09397093951702118, 0.1428760588169098, 0.08571921288967133, 0.2478504776954651, 0.12005335837602615, -0.22748184204101562, 0.055358242243528366, 0.12515293061733246, -0.14365963637828827, 0.10365243256092072, 0.07344598323106766, 0.005470725707709789, -0.18886831402778625, -0.06843198090791702, -0.06121627986431122, 0.1053021252155304, -0.08522345870733261, -0.0776243582367897, 0.09323626756668091, -0.05086790770292282, 0.24641476571559906, -0.028281206265091896, 0.06174173951148987, -0.026681531220674515, -0.1389324963092804, -0.01723906397819519, 0.060955192893743515, 0.05258452147245407, -0.024835573509335518, -0.25895482301712036, 0.13646544516086578, 0.048650871962308884, 0.025074828416109085, 0.004106190986931324, -0.05691491439938545, 0.016934165731072426, 0.1511998474597931, 0.020012924447655678, 0.13717477023601532, 0.027723990380764008, 0.0706823319196701, -0.006239562761038542, -0.10560829937458038, -0.04169593006372452, 0.061916545033454895, -0.08518962562084198, -0.06641357392072678, 0.011197872459888458, -0.06935211271047592, -0.11783787608146667, -0.12166737765073776, -0.026334572583436966, -0.02980303019285202, -0.07444227486848831, 0.02368103712797165, 0.06536602973937988, -0.06702698022127151, -0.0023908785078674555, 0.007125476840883493, -0.011537045240402222, 0.16434046626091003, 0.011393417604267597, -0.007796820718795061, 0.1328643560409546, -0.11533161997795105, 0.12461213022470474, 0.049438029527664185, -0.024806302040815353, -0.04662557691335678, 0.0014137453399598598, -0.057529181241989136, 0.029044216498732567, -0.04390640929341316, 0.02774495631456375, 0.20111067593097687, 0.02772962674498558, 0.11389166116714478, -0.0656520202755928, 0.04385066404938698, -0.007961965166032314, -0.009693224914371967, 0.018563594669103622, 0.07608018070459366, 0.07813210040330887, -0.1324140727519989, 0.02262016013264656, 0.22455167770385742, 0.1385764330625534, 0.18313980102539062, -0.010877152904868126, 0.06325667351484299, -0.04875868931412697, 0.027505528181791306, 0.024100203067064285, 0.10314226150512695, -0.10732068121433258, -0.0322517491877079, -0.025407759472727776, 0.023599207401275635, -0.08197105675935745, -0.1055799350142479, -0.090115025639534, 0.01222382951527834, -0.03125503659248352, -0.15570329129695892, 0.13300658762454987, -0.10451057553291321, 0.01802753657102585, 0.04692702740430832, -0.22163605690002441, 0.11530312895774841, 0.014291439205408096, -0.10303618758916855, 0.11281087249517441, -0.12051989883184433, -0.08699832111597061, -0.05777236074209213, -0.18658851087093353, 0.05280197039246559, 0.04673841595649719, 0.05166793242096901, -0.18521739542484283, 0.024835903197526932, 0.05545609071850777, 0.13426995277404785, -0.09743253141641617, -0.07142634689807892, -0.15038461983203888, 0.016068490222096443, -0.033661190420389175, -0.16029728949069977, -0.005609163548797369, -0.032781440764665604, -0.18849676847457886, -0.04539939761161804, -0.15086813271045685, -0.034627582877874374, 0.20464378595352173, 0.026907702907919884, 0.09480511397123337, -0.07926445454359055, 0.3802889585494995, -0.042039383202791214, -0.06146497279405594, -0.01321389526128769, -0.07072482258081436, 0.02512686513364315, 0.13271741569042206, 0.0036099457647651434, -0.017886579036712646, -0.0037857077550143003, 0.0024592927657067776, -0.06234965845942497, -0.13400450348854065, 0.0028710351325571537, 0.03905198723077774, 0.1874423623085022, 0.004639793653041124, 0.06659388542175293, 0.03133883699774742, 0.057546284049749374, 0.07748064398765564, 0.030926106497645378, 0.0011591583024710417, -0.01591806672513485, 0.06604493409395218, -0.11684755235910416, 0.042466625571250916, -0.030429253354668617, -0.10143838077783585, -0.013183288276195526, 0.07950251549482346, 0.12755028903484344, 0.17849206924438477, -0.04790908098220825, 0.17489230632781982, 0.13580141961574554, 0.16576050221920013, 0.049315933138132095, -0.020801831036806107, -0.08773037046194077, -0.06118565797805786, 0.004774159751832485, -0.031952597200870514, 0.04869702458381653, 0.3231290578842163, 0.037619613111019135, -0.09036035090684891, 0.11149907857179642, 0.009480619803071022, 0.05359881371259689, 0.022797370329499245, -0.11162138730287552, 0.11170321702957153, 0.07968773692846298, -0.06341761350631714, -0.07602835446596146, 0.16758501529693604, -0.1109386757016182, -0.26646625995635986, -0.11410990357398987, -0.012305386364459991, 0.07903840392827988, 0.005651174578815699, 0.05498376116156578, -0.11829282343387604, -0.16034497320652008, -0.034191906452178955, 0.1335442066192627, -0.3077351450920105, 0.2065143585205078, -0.0198091771453619, 0.06707923114299774, -0.039657969027757645, -0.07026876509189606, 0.09694647043943405, 0.13174086809158325, 0.29124146699905396, 0.01396956667304039, 0.04841272905468941, -0.15176129341125488, -0.0976925864815712, 0.0018439020495861769, 0.015482662245631218, -0.02563396655023098, 0.028520405292510986, -0.0540912002325058, 0.008404579944908619, -0.018086453899741173, 0.2102297693490982, -0.11316607892513275, 0.004344627261161804, -0.06968966871500015, -0.11707738786935806, 0.19409789144992828, -0.07178345322608948, -0.04543264955282211, -0.14959357678890228, -0.15512511134147644, -0.004174166824668646, -0.02413962036371231, -0.019664527848362923, -0.17603960633277893, -0.18804074823856354, -0.05204557999968529, -0.005645004566758871, -0.003464865731075406, 0.05867868289351463, -0.07517234236001968, -0.04805335775017738, 0.1009904220700264, -0.07743175327777863, -0.056063808500766754, -0.1103200614452362, 0.1391381323337555, 0.06248528137803078, 0.16743235290050507, 0.05907081440091133, 0.0006117874872870743, 0.11471151560544968, -0.02913086675107479, 0.11103474348783493, -0.11291708797216415, -0.17145049571990967, -0.08334989100694656, -0.018775060772895813, 0.09519003331661224, -0.04789286106824875, 0.0028788831550627947, 0.2550160884857178, 0.14880181849002838, -0.0897710770368576, 0.27680760622024536, 0.04414956644177437, -0.09375058114528656, -0.18432219326496124, -0.15961645543575287, 0.03759992495179176, 0.060025621205568314, 0.13095876574516296, -0.057205069810152054, -0.08483537286520004, -0.08492398262023926, -0.07478608191013336, -0.13140805065631866, -0.24232175946235657, -0.030598774552345276, 0.22874866425991058, 0.08656918257474899, 0.08219650387763977, -0.012482990510761738, -0.01186054851859808, 0.00526038184762001, 0.02680150233209133, 0.12018456310033798, -0.13341329991817474, 0.11107480525970459, 0.022198403254151344, 0.044267985969781876, 0.009712530300021172, 0.07929777354001999, 0.03375575691461563, -0.003218587953597307, -0.0006439819699153304, -0.0988350659608841, -0.2596651017665863, 0.0816885456442833, -0.01623627357184887, -0.09960969537496567, 0.014988959766924381, 0.02061903104186058, -0.2089255303144455, 0.011128270998597145, -0.019883770495653152, -0.03150356933474541, -0.06483490765094757, -0.10664787143468857, -0.056551624089479446, 0.04928823933005333, 0.10853826254606247, 0.011660109274089336, 0.05354316532611847, -0.0404130220413208, 0.07917837053537369, 0.0826287642121315, 0.15132710337638855, 0.06795957684516907, -0.190711110830307, -0.10953907668590546, -0.0414445661008358, 0.12121522426605225, -0.12505418062210083, 0.036917757242918015, 0.053161121904850006, -0.016534561291337013, 0.14621229469776154, 0.1070784479379654, -0.07452095299959183, 0.11915595084428787, 0.08904775977134705, -0.04094788804650307, -0.23367151618003845, -0.07120766490697861, 0.11133213341236115, 0.07195597887039185, -0.03961895406246185, 0.018120890483260155, -0.04960581287741661, -0.013980977237224579, 0.048759616911411285, -0.0538676381111145, -0.07230538129806519, 0.004421027842909098, 0.1247575581073761, 0.1029362753033638, -0.04655474051833153, 0.01296416949480772, 0.037371400743722916, 0.003788623260334134, 0.04730486497282982, 0.0407949760556221, -0.08269952982664108, -0.04124005511403084, 0.02782733179628849, 0.37552911043167114, -0.010165480896830559, -0.020456433296203613, 0.018555615097284317, -0.19949445128440857, 0.09135842323303223, 0.13205479085445404, 0.04697350412607193, 0.004247748292982578, -0.08139242231845856, 0.026877427473664284, -0.010625290684401989, 0.09936143457889557, -0.07806670665740967, -0.05493134260177612, -0.21631066501140594, -0.025010565295815468, 0.017490221187472343, 0.24077683687210083, -0.08458559215068817, -0.12801732122898102, -0.20628872513771057, 0.13128381967544556, -0.11333390325307846, -0.03695881739258766, -0.024473199620842934, 0.03926658630371094, -0.01989821158349514, 0.06291737407445908, -0.0710630789399147, 0.006373001262545586, -0.11024709790945053, 0.055267609655857086, 0.04204455390572548, 0.1229788213968277, 0.014207782223820686, 0.02016810141503811, 0.05822525918483734, -0.01837925612926483, 0.07173580676317215, -0.06203491613268852, -0.04550490900874138, 0.14224006235599518, -0.020255116745829582, -0.04152837023139, -0.0483345128595829, -0.036874305456876755, 0.11981741338968277, -0.05059147998690605, -0.007141099311411381, -0.054929375648498535, -0.06906463205814362, 0.03462086617946625, -0.009175732731819153, -0.008798843249678612, 0.06801853328943253, 0.04024988040328026, -0.026994358748197556, 0.005263668950647116, 0.03447828069329262, -0.10330043733119965, -0.04955084249377251, 0.16955432295799255, -0.0749620869755745, 0.10274054110050201, -0.031069839373230934, 0.018015999346971512, 0.005847334861755371, -0.022399673238396645, -0.015360680408775806, -0.1457086056470871, -0.06137600541114807, -0.09489979594945908, 0.11565322428941727, 0.08146517723798752, 0.03358805552124977, 0.04274565726518631, 0.019532648846507072, -0.04414922371506691, -0.038583990186452866, 0.12961317598819733, 0.08133101463317871, 0.012996876612305641, 0.01137041300535202, 0.01941833831369877, -0.020302120596170425, 0.0028480992186814547, -0.01250747125595808, -0.07239153981208801, -0.05874783173203468, 0.09400010108947754, 0.1600283533334732, -0.06127211079001427, -0.13325586915016174, -0.020593497902154922, 0.04988488554954529, 0.0014717020094394684, -0.08777432143688202, 0.04833676666021347, 0.15805292129516602, -0.05623878911137581, 0.03216489031910896, -0.09984751045703888, -0.07263360917568207, -0.16060975193977356, -0.10029061883687973, -0.06092562898993492, -0.28350353240966797, 0.09752398729324341, 0.006392303854227066, -0.014731393195688725, 0.059529416263103485, 0.051305368542671204, -0.052508849650621414, 0.07068239152431488, -0.18146829307079315, -0.007054794579744339, 0.03497592359781265, -0.13212306797504425, 0.02475893869996071, -0.2378365397453308, 0.10198072344064713, -0.04623803123831749, -0.1519704908132553, -0.04004510119557381, 0.0641569048166275, -0.09540136158466339, -0.01822364516556263, -0.0475153923034668, -0.01922670193016529, 0.01624443754553795, -0.009348669089376926, -0.031147832050919533, 0.13716529309749603, 0.02827494591474533, -0.03268734738230705, 0.005254602525383234, 0.0223685409873724, 0.03955082967877388, -0.0969657450914383, -0.05986930429935455, 0.08311155438423157, -0.031056145206093788, 0.14728976786136627, 0.000341245875461027, 0.04181376099586487, -0.06758682429790497, 0.2593761384487152, 0.2023983597755432, -0.12479214370250702, 0.008118697442114353, -0.021801479160785675, 0.012670028023421764, -0.041751839220523834, 0.13110700249671936, 0.013386172242462635, 0.12186761200428009, -0.17513342201709747, -0.01036517322063446, -0.0818324014544487, -0.04501292482018471, 0.06702108681201935, 0.14714950323104858, 0.15742522478103638, 0.03436789661645889, -0.07328428328037262, 0.06722653657197952, -0.30119743943214417, 0.20540550351142883, -0.1346001923084259, -0.01498429011553526, -0.040251150727272034, -0.058389630168676376, 0.061147745698690414, 0.11309876292943954, 0.10832664370536804, -0.021150551736354828, -0.0905047357082367, -0.04486766457557678, -0.039378076791763306, -0.13019338250160217, -0.02718670479953289, 0.1654091775417328, 0.06799814850091934, 0.31520840525627136, -0.017577875405550003, 0.07702425122261047, 0.034410297870635986, 0.06451138854026794, 0.004519328009337187, 0.09537279605865479, 0.07960964739322662, -0.06345855444669724, -0.07373003661632538, -0.001637450186535716, 0.05033271387219429, 0.14567798376083374, -0.03826142102479935, -0.18691548705101013, 0.15858715772628784, 0.07192251086235046, -0.13762691617012024, -0.05777517706155777, 0.08409425616264343, -0.0739973932504654, 0.0550808347761631, 0.08115427941083908, 0.015876613557338715, -0.017793258652091026, -0.004664506763219833, 0.06074233725667, 0.024694660678505898, -0.02343848906457424, 0.003570882137864828, -0.08337053656578064, -0.04151543974876404, 0.07267895340919495, -0.0844460055232048, -0.20546193420886993, -0.0957019031047821, -0.07551700621843338, 0.030557552352547646, -0.0649830624461174, 0.12575586140155792, 0.1717868149280548, 0.0593598335981369, -0.03307248651981354, -0.10721943527460098, -0.035562749952077866, 0.07602505385875702, -0.044773899018764496, -0.09409699589014053 ]
null
null
transformers
## **BigMaid-20B-v1.0** [exllamav2](https://github.com/turboderp/exllamav2) quant for [TeeZee/BigMaid-20B-v1.0](https://huggingface.co/TeeZee/BigMaid-20B-v1.0) Should run on 12 GB of VRAM cards in webui with context length set to 4096, ExLlamav2_HF loader and cache_8bit=True All comments are greatly appreciated, download, test and if you appreciate my work, consider buying me my fuel: <a href="https://www.buymeacoffee.com/TeeZee" target="_blank"><img src="https://cd
{"license": "apache-2.0", "tags": ["merge", "not-for-all-audiences"]}
text-generation
TeeZee/BigMaid_20B_v1.0-bpw4-h6-exl2
[ "transformers", "safetensors", "llama", "text-generation", "merge", "not-for-all-audiences", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T12:53:04+00:00
[]
[]
TAGS #transformers #safetensors #llama #text-generation #merge #not-for-all-audiences #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
## BigMaid-20B-v1.0 exllamav2 quant for TeeZee/BigMaid-20B-v1.0 Should run on 12 GB of VRAM cards in webui with context length set to 4096, ExLlamav2_HF loader and cache_8bit=True All comments are greatly appreciated, download, test and if you appreciate my work, consider buying me my fuel: <a href="URL target="_blank"><img src="https://cd
[ "## BigMaid-20B-v1.0\n\nexllamav2 quant for TeeZee/BigMaid-20B-v1.0\n\nShould run on 12 GB of VRAM cards in webui with context length set to 4096, ExLlamav2_HF loader\nand cache_8bit=True\n\nAll comments are greatly appreciated, download, test and if you appreciate my work, consider buying me my fuel:\n<a href=\"URL target=\"_blank\"><img src=\"https://cd" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #merge #not-for-all-audiences #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "## BigMaid-20B-v1.0\n\nexllamav2 quant for TeeZee/BigMaid-20B-v1.0\n\nShould run on 12 GB of VRAM cards in webui with context length set to 4096, ExLlamav2_HF loader\nand cache_8bit=True\n\nAll comments are greatly appreciated, download, test and if you appreciate my work, consider buying me my fuel:\n<a href=\"URL target=\"_blank\"><img src=\"https://cd" ]
[ 67, 110 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #merge #not-for-all-audiences #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n## BigMaid-20B-v1.0\n\nexllamav2 quant for TeeZee/BigMaid-20B-v1.0\n\nShould run on 12 GB of VRAM cards in webui with context length set to 4096, ExLlamav2_HF loader\nand cache_8bit=True\n\nAll comments are greatly appreciated, download, test and if you appreciate my work, consider buying me my fuel:\n<a href=\"URL target=\"_blank\"><img src=\"https://cd" ]
[ -0.03086857497692108, -0.1447758823633194, -0.0019936945755034685, 0.10542882978916168, 0.07702162861824036, 0.04051894694566727, 0.04087613895535469, 0.0975964292883873, 0.003012771485373378, -0.021538585424423218, 0.112208791077137, 0.027190228924155235, 0.009508971124887466, 0.12200907617807388, 0.007225253153592348, -0.06341268867254257, 0.012212902307510376, 0.03838355466723442, 0.10306378453969955, 0.11232367902994156, 0.16298750042915344, -0.1271078884601593, 0.14455606043338776, -0.025580087676644325, -0.12361521273851395, 0.031990598887205124, -0.039260491728782654, -0.044464949518442154, 0.0555705651640892, 0.026179980486631393, 0.0530419647693634, 0.05277928337454796, -0.005383700132369995, -0.030284518375992775, 0.0630212128162384, 0.029281744733452797, -0.06173847243189812, 0.0263481717556715, 0.022190187126398087, 0.003579223994165659, 0.07899711281061172, -0.05740936100482941, -0.06163541227579117, 0.08043969422578812, 0.013072626665234566, -0.19126155972480774, -0.07670528441667557, -0.009995675645768642, 0.03625565022230148, 0.05065762624144554, -0.00018835547962225974, 0.1312563717365265, -0.07346667349338531, 0.07904618978500366, 0.2572936415672302, -0.1501781940460205, -0.01960323378443718, 0.0887899100780487, 0.06820915639400482, 0.08900362998247147, -0.03540613874793053, 0.002158067189157009, 0.0737103670835495, 0.01722322590649128, 0.12263929098844528, -0.008120213635265827, -0.011086852289736271, -0.02283383533358574, -0.048825547099113464, 0.03302421048283577, 0.3050738275051117, 0.08136112987995148, -0.04776030406355858, -0.05581780523061752, -0.08087530732154846, -0.17127226293087006, -0.02799021266400814, -0.01572120189666748, 0.04114021733403206, -0.03849508985877037, 0.029902871698141098, -0.0200502946972847, -0.12744669616222382, -0.06157156825065613, -0.14158079028129578, 0.006241821683943272, 0.016966108232736588, 0.03453509882092476, -0.04403673857450485, 0.08707903325557709, 0.02910310961306095, -0.13478504121303558, -0.1064833402633667, -0.038812942802906036, 0.05165325477719307, -0.03517095372080803, -0.03443045914173126, 0.010346975177526474, 0.05694536492228508, 0.044339150190353394, 0.07499270141124725, 0.020828869193792343, 0.029161356389522552, 0.023770326748490334, -0.0150283332914114, -0.049123119562864304, -0.0515705905854702, -0.11986851692199707, 0.10747005045413971, -0.010675068013370037, 0.13174594938755035, 0.03454294428229332, -0.11533306539058685, -0.0033732112497091293, 0.036640606820583344, 0.036077823489904404, -0.04106338322162628, -0.0020772749558091164, -0.05172546207904816, 0.012803560122847557, 0.17015641927719116, -0.06634610891342163, -0.004002898931503296, 0.03888871148228645, -0.046694934368133545, 0.1454860121011734, 0.013930562883615494, -0.014376731589436531, 0.02694939263164997, 0.013886826112866402, -0.05584679916501045, -0.025819607079029083, -0.06818324327468872, -0.03079831227660179, 0.12926548719406128, -0.008652638643980026, 0.04157373681664467, -0.2225182205438614, -0.21617747843265533, 0.07939867675304413, 0.008460722863674164, -0.06588394939899445, 0.007650522980839014, 0.020039597526192665, -0.04171399772167206, 0.014686495997011662, -0.007112259976565838, 0.1971503049135208, -0.0821329727768898, 0.0660240650177002, 0.0931338295340538, 0.12293986231088638, -0.07932768017053604, 0.03618518263101578, -0.10027280449867249, -0.02298460528254509, 0.0028370311483740807, -0.036586515605449677, -0.017944157123565674, 0.06893512606620789, -0.08633439987897873, -0.07319866865873337, -0.07811231166124344, 0.026454295963048935, -0.009764831513166428, 0.10138160735368729, -0.1652192622423172, -0.018057961016893387, 0.11116648465394974, -0.07834508270025253, -0.21530935168266296, 0.1526349037885666, 0.006781549192965031, 0.0508817695081234, 0.05233502388000488, -0.0947718545794487, 0.1172885149717331, -0.14726980030536652, -0.09858743846416473, 0.07462381571531296, -0.030695876106619835, -0.19333794713020325, 0.12729008495807648, 0.1284526139497757, 0.012697512283921242, 0.017992299050092697, 0.14550790190696716, 0.04223336651921272, 0.012822341173887253, -0.0736418142914772, -0.1426936686038971, -0.08274288475513458, 0.009869968518614769, 0.0031049237586557865, 0.015571183525025845, -0.11589299887418747, -0.0981232151389122, -0.04582381248474121, 0.14099597930908203, 0.045086540281772614, -0.0013525434769690037, -0.07441779226064682, 0.0857255831360817, -0.15387165546417236, 0.03195701912045479, 0.016706684604287148, -0.057582661509513855, 0.00039163650944828987, -0.057173337787389755, -0.04494796320796013, 0.1500624418258667, 0.07768018543720245, -0.014035812579095364, -0.0826495811343193, 0.009393042884767056, 0.056503795087337494, 0.004593724850565195, -0.01751839742064476, -0.06905421614646912, 0.03980381414294243, -0.016951337456703186, 0.09539611637592316, -0.0052717300131917, 0.006035428494215012, 0.228496253490448, 0.07389570772647858, 0.033659160137176514, 0.03091450408101082, 0.02753472700715065, -0.044998083263635635, -0.07871656864881516, -0.014969885349273682, 0.02651185542345047, -0.0352865569293499, -0.04857741296291351, 0.0904405415058136, -0.07838491350412369, 0.2070811539888382, 0.21327485144138336, 0.011233345605432987, 0.03360303118824959, 0.03646222874522209, -0.0021521796006709337, -0.04699209704995155, -0.04922698065638542, -0.13227824866771698, -0.12322535365819931, 0.009004364721477032, 0.1083168238401413, -0.08812566846609116, -0.07932313531637192, 0.055082641541957855, 0.010456204414367676, -0.019166208803653717, 0.06603990495204926, 0.08148934692144394, -0.11014338582754135, 0.11944574117660522, 0.2854779064655304, -0.08507300168275833, 0.0832214429974556, 0.003261054400354624, -0.04708147794008255, -0.009677717462182045, -0.027769507840275764, -0.018808841705322266, 0.10522811859846115, -0.05756383016705513, 0.07370738685131073, 0.08366803824901581, -0.0010252518113702536, 0.0005467954906634986, -0.04650996997952461, 0.010476103983819485, 0.03870102018117905, -0.05059411749243736, 0.04675740376114845, 0.11436368525028229, -0.04276275262236595, 0.04685451462864876, -0.0424668975174427, -0.0010947947157546878, 0.09804357588291168, 0.027475450187921524, -0.039285365492105484, 0.11582384258508682, -0.04915079474449158, -0.20216228067874908, -0.1746245175600052, -0.06756167858839035, -0.052504897117614746, 0.0008861200185492635, 0.1062847375869751, -0.10221049189567566, -0.08005576580762863, -0.06582114845514297, -0.039419807493686676, 0.06276386976242065, 0.035136427730321884, -0.025943202897906303, 0.03697013482451439, 0.02108227089047432, -0.12328886240720749, -0.00504369055852294, 0.10289394855499268, -0.04343673214316368, 0.08816799521446228, 0.051936548203229904, 0.10586504638195038, 0.04951735585927963, 0.002388689434155822, -0.021729961037635803, 0.008442177437245846, 0.0742344930768013, -0.06754063069820404, 0.12929943203926086, 0.2624383866786957, 0.09250625222921371, 0.01907780021429062, 0.04117968678474426, 0.038843076676130295, -0.03175429254770279, 0.062148116528987885, 0.00950667541474104, -0.12049959599971771, -0.17040377855300903, -0.04416019469499588, -0.07255200296640396, 0.014221656136214733, 0.020687496289610863, 0.04482714459300041, -0.03410368412733078, 0.15536925196647644, -0.07839404791593552, 0.05585337430238724, -0.09175895154476166, 0.03503021225333214, 0.10131311416625977, 0.009205015376210213, 0.09596172720193863, -0.12302696704864502, -0.017815686762332916, 0.1227496787905693, 0.040782470256090164, -0.008504177443683147, -0.016887880861759186, 0.21118147671222687, 0.04048759490251541, 0.049506429582834244, 0.023198243230581284, 0.13621880114078522, -0.013709048740565777, -0.05374591797590256, -0.04657222703099251, -0.06971611082553864, 0.015592575073242188, 0.05092020332813263, -0.07163701206445694, 0.09876998513936996, -0.0116141801699996, 0.07760719954967499, 0.08160737156867981, 0.10215489566326141, 0.007436785381287336, -0.26340723037719727, -0.011866776272654533, 0.11556284874677658, -0.021869629621505737, -0.03210194408893585, -0.0029525819700211287, 0.06768859177827835, -0.014825459569692612, 0.053692758083343506, -0.0009139904868789017, 0.06150215119123459, 0.00045297216274775565, 0.02603006921708584, -0.011685664765536785, 0.12240180373191833, 0.02750280499458313, 0.11823644489049911, -0.3291490077972412, 0.019861003383994102, 0.0390898659825325, 0.02541166916489601, -0.13523557782173157, -0.015262809582054615, 0.027906080707907677, 0.008174042217433453, 0.06310984492301941, -0.02360612340271473, -0.004683185834437609, 0.051160141825675964, -0.10809971392154694, 0.052094295620918274, 0.004936201963573694, 0.031338486820459366, 0.043006058782339096, -0.03716215491294861, -0.027562620118260384, 0.011382827535271645, 0.0957147404551506, -0.06883572041988373, -0.08456364274024963, 0.02585407719016075, 0.1631709188222885, -0.016913438215851784, -0.07738400250673294, -0.055764175951480865, -0.15980137884616852, 0.10724616795778275, -0.2031521052122116, -0.07888223230838776, -0.0497707724571228, -0.09779851138591766, 0.029300382360816002, -0.0552678145468235, 0.050128620117902756, -0.03317864611744881, 0.13038906455039978, -0.01793706603348255, -0.05510532110929489, 0.07856538146734238, -0.13562332093715668, -0.11844882369041443, -0.0067305788397789, 0.11694575846195221, -0.09735868126153946, 0.05630623921751976, 0.0028507017996162176, -0.03273320943117142, -0.12927037477493286, -0.09925904124975204, -0.01959848590195179, -0.012320341542363167, -0.02163466066122055, 0.003375197993591428, -0.011674425564706326, -0.007449105381965637, 0.12738826870918274, -0.11954569816589355, 0.13205178081989288, 0.17234328389167786, -0.04910198599100113, 0.06329434365034103, 0.10672539472579956, 0.03402797877788544, -0.23072503507137299, -0.05936724692583084, -0.0994940921664238, -0.011103183962404728, -0.011414481326937675, -0.036068372428417206, 0.1728406697511673, 0.11818048357963562, -0.043758004903793335, 0.16595056653022766, -0.17128244042396545, -0.10963547229766846, 0.02573508583009243, 0.020824434235692024, 0.28946653008461, -0.09925122559070587, 0.001966227078810334, -0.14309027791023254, -0.20263294875621796, 0.09421183913946152, -0.25966036319732666, 0.10776428878307343, -0.06828130036592484, 0.08025701344013214, -0.04233204945921898, -0.07893561571836472, 0.10677775740623474, -0.0635785311460495, 0.054515812546014786, -0.11989377439022064, 0.09509767591953278, 0.0926983654499054, -0.045575905591249466, 0.12416405230760574, -0.24647113680839539, -0.037669017910957336, -0.08416203409433365, -0.0197101142257452, -0.027469679713249207, 0.06471997499465942, -0.032663874328136444, -0.05078735575079918, -0.061348266899585724, -0.006445660721510649, -0.007264697924256325, -0.09818786382675171, 0.07647041976451874, -0.0007562619866803288, -0.03269724175333977, 0.20135053992271423, 0.1284598410129547, -0.17798537015914917, -0.01689240150153637, -0.029304003342986107, -0.053989045321941376, 0.058455631136894226, -0.15536223351955414, 0.08460571616888046, 0.011781612411141396, -0.057459890842437744, 0.021231627091765404, 0.03637458384037018, -0.03634033352136612, 0.05356581136584282, 0.15200427174568176, -0.09964215755462646, -0.09190908074378967, -0.05509015545248985, 0.041262056678533554, -0.12524636089801788, -0.00008624166366644204, 0.10647718608379364, -0.032085224986076355, 0.03817269951105118, 0.03247135505080223, 0.04983726888895035, -0.05739397928118706, 0.11855045706033707, 0.08264193683862686, 0.06627465039491653, -0.11558876186609268, 0.057969652116298676, -0.01420633401721716, -0.08264860510826111, -0.01286224089562893, 0.05758122727274895, -0.07785461843013763, -0.1430719792842865, 0.04829569533467293, -0.059979066252708435, -0.04934954643249512, -0.1033078208565712, -0.08772493153810501, -0.13570883870124817, 0.018486054614186287, 0.01609650067985058, 0.12249951809644699, 0.045769453048706055, 0.07965002208948135, -0.004655075259506702, -0.09370536357164383, 0.12950608134269714, -0.020754938945174217, 0.10824806243181229, -0.1872485727071762, -0.024881860241293907, -0.048818204551935196, 0.0777997076511383, -0.05788220465183258, 0.09469703584909439, -0.01878933608531952, -0.01759256236255169, -0.20179490745067596, 0.012143396772444248, -0.03174309805035591, 0.006739079020917416, 0.007758296560496092, 0.03626884147524834, -0.07018274068832397, 0.047995276749134064, -0.03556642308831215, 0.02403695322573185, 0.03629819676280022, 0.04959222301840782, -0.0685124397277832, -0.007713702041655779, 0.03222782537341118, -0.010537509806454182, 0.06384525448083878, -0.03179657459259033, -0.04215312749147415, -0.016287492588162422, -0.14459864795207977, 0.005323308520019054, 0.06568551808595657, 0.06973722577095032, 0.0315629281103611, -0.042991600930690765, 0.05530877038836479, 0.036804597824811935, 0.038090962916612625, 0.022913919761776924, 0.05623922869563103, -0.11547341197729111, 0.055767692625522614, -0.07920152693986893, 0.009215084835886955, -0.03486498072743416, -0.017659801989793777, 0.04325209558010101, -0.03984689339995384, 0.20734558999538422, -0.019202932715415955, -0.04098660498857498, -0.09617944806814194, 0.0791139230132103, -0.045664671808481216, -0.1381334513425827, -0.10572866350412369, -0.041888944804668427, -0.002938412129878998, -0.008491612039506435, 0.1752873808145523, -0.03149563819169998, -0.046579740941524506, 0.0609765462577343, 0.02947767823934555, 0.01663433201611042, -0.0444706454873085, 0.18192033469676971, 0.0010390195529907942, 0.015780067071318626, -0.06922672688961029, 0.021389953792095184, 0.13391228020191193, -0.09335644543170929, 0.08198942244052887, 0.14462853968143463, -0.09977008402347565, 0.1308610886335373, 0.012704881839454174, -0.04934530705213547, 0.015094214119017124, -0.040364354848861694, -0.05445102974772453, 0.049511171877384186, -0.043757058680057526, 0.06253766268491745, 0.11881742626428604, -0.04024074971675873, -0.06775709241628647, -0.03950170427560806, 0.0280462633818388, -0.09631425887346268, 0.03919079154729843, -0.15871039032936096, -0.10911434888839722, -0.012846876867115498, -0.024262841790914536, -0.05466407537460327, 0.02312476933002472, 0.06730963289737701, -0.04881250858306885, 0.22646057605743408, 0.004312878008931875, -0.0015217956388369203, 0.019664226099848747, 0.0028295484371483326, -0.09723927825689316, 0.06635871529579163, -0.05700867623090744, -0.00967654399573803, -0.052568189799785614, -0.028942294418811798, 0.035237882286310196, -0.023770160973072052, -0.012760588899254799, -0.05316664278507233, -0.07117076218128204, -0.0695469081401825, 0.038776740431785583, 0.03167654946446419, 0.038288701325654984, -0.009400990791618824, 0.010081816464662552, 0.019976302981376648, 0.08031909167766571, -0.02972210757434368, -0.08473207801580429, -0.013934684917330742, -0.007542442996054888, 0.01153919193893671, 0.05493660271167755, -0.0045727575197815895, -0.05399951711297035, 0.004141886718571186, 0.20503942668437958, 0.1520390510559082, -0.029471294954419136, 0.03731531649827957, -0.005375796463340521, 0.007989897392690182, -0.04712508246302605, 0.07289183884859085, 0.09201128780841827, 0.0859600380063057, -0.026823515072464943, 0.14146053791046143, -0.004005478695034981, 0.0029589030891656876, -0.08133481442928314, 0.044541940093040466, -0.05200900882482529, 0.012253363616764545, 0.03416115790605545, -0.040243808180093765, -0.02562524937093258, -0.007976632565259933, 0.015008551068603992, -0.02578970044851303, -0.04520547017455101, -0.006438846699893475, 0.11821548640727997, 0.061324942857027054, -0.002959526376798749, -0.05279091000556946, -0.01863248646259308, 0.04213401675224304, 0.009761001914739609, -0.16584260761737823, -0.025416744872927666, 0.06869073212146759, -0.052212007343769073, 0.12647704780101776, -0.038240984082221985, 0.08746788650751114, 0.12992292642593384, -0.06817896664142609, -0.16083431243896484, 0.12436281144618988, 0.013548215851187706, -0.14798440039157867, 0.10179100930690765, 0.062391042709350586, -0.03334783390164375, 0.0013103899545967579, 0.020697658881545067, 0.012431366369128227, -0.062319766730070114, 0.08523265272378922, 0.017617668956518173, -0.07684873789548874, 0.0767659991979599, -0.04100578650832176, 0.100135937333107, 0.0759064108133316, -0.07691631466150284, -0.051130637526512146, -0.07180438935756683, 0.122248075902462, 0.039131276309490204, -0.0496055968105793, 0.026112454012036324, -0.13212986290454865, 0.013859032653272152, 0.11243725568056107, 0.05954710394144058, -0.29210883378982544, -0.04799085110425949, -0.07843754440546036, -0.02603289484977722, -0.12311940640211105, 0.0076169599778950214, 0.1865011602640152, 0.04052874818444252, -0.02187981642782688, -0.0674867331981659, -0.10410450398921967, 0.054173316806554794, -0.1785273402929306, -0.09727910906076431 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
null
Sakil/mistral_7b_finetuned
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
2024-02-10T12:57:23+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 31, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.06646376848220825, 0.2168014943599701, -0.00225935154594481, 0.023818302899599075, 0.1271018385887146, -0.001635765191167593, 0.04218708351254463, 0.13324736058712006, -0.020175931975245476, 0.11144465953111649, 0.046588581055402756, 0.09377603232860565, 0.09928803145885468, 0.18404334783554077, 0.04859916493296623, -0.2059975117444992, 0.007056170143187046, -0.09090408682823181, 0.014076028019189835, 0.1116579994559288, 0.13719257712364197, -0.10291384905576706, 0.08272874355316162, -0.04045208916068077, -0.02019004337489605, 0.00012576708104461432, -0.09259183704853058, -0.07032395154237747, 0.06885425746440887, 0.06264153122901917, 0.051234472543001175, 0.001456156256608665, 0.09140396863222122, -0.2864592671394348, 0.017265573143959045, 0.08406311273574829, 0.0027674848679453135, 0.06290827691555023, 0.07236549258232117, -0.07389893382787704, 0.11328595131635666, -0.08021481335163116, 0.13019037246704102, 0.08625296503305435, -0.062064990401268005, -0.23071379959583282, -0.07525765895843506, 0.0963398814201355, 0.12251301854848862, 0.06215599179267883, -0.022921854630112648, 0.15455181896686554, -0.06248689442873001, 0.012971068732440472, 0.1294165402650833, -0.11526761949062347, -0.05572471022605896, 0.061741601675748825, 0.11775490641593933, 0.10740239918231964, -0.14110268652439117, -0.0017287094378843904, 0.04900608956813812, 0.029121357947587967, 0.08589313924312592, 0.022661056369543076, 0.12003941088914871, 0.04652795568108559, -0.13695219159126282, -0.04037507623434067, 0.12011898308992386, 0.038862764835357666, -0.06446044892072678, -0.2168138176202774, -0.006778308190405369, -0.0601806715130806, -0.014732478186488152, -0.07019448280334473, 0.039128515869379044, -0.02470310963690281, 0.07317749410867691, -0.04465159401297569, -0.1063927412033081, -0.0421026237308979, 0.0892222449183464, 0.07748593389987946, 0.011527054943144321, -0.02519804798066616, 0.04627908393740654, 0.13455867767333984, 0.05402068421244621, -0.10399353504180908, -0.07017925381660461, -0.06942764669656754, -0.09420394152402878, -0.04035796597599983, 0.056760527193546295, 0.031942449510097504, 0.02665667235851288, 0.22703726589679718, 0.016653569415211678, 0.04155244305729866, 0.0224777739495039, 0.01032855175435543, 0.043662428855895996, 0.0955500528216362, -0.05303520709276199, -0.15660029649734497, -0.04072032496333122, 0.09077946096658707, -0.0027527001220732927, -0.036689214408397675, -0.03966725245118141, 0.03849169611930847, 0.06843466311693192, 0.13122352957725525, 0.07552056759595871, -0.017929591238498688, -0.04813180863857269, -0.030096933245658875, 0.23523783683776855, -0.1493375599384308, 0.04426715523004532, -0.02271856553852558, -0.01804111897945404, -0.03908449783921242, 0.03597262129187584, 0.022118929773569107, -0.000004518366949923802, 0.09706240892410278, -0.058981191366910934, -0.05378659814596176, -0.10168042778968811, -0.03272576630115509, 0.04088849574327469, -0.013975566253066063, -0.010589460842311382, -0.09025166928768158, -0.09490354359149933, -0.04766594246029854, 0.05537205561995506, -0.05123869329690933, -0.03770573064684868, 0.009465423412621021, -0.08151785284280777, -0.005444355774670839, -0.005417742300778627, 0.10699385404586792, -0.03222226724028587, 0.04445803165435791, -0.027600755915045738, 0.05225523188710213, 0.09919606149196625, 0.031576547771692276, -0.0773419588804245, 0.0561848059296608, -0.22559374570846558, 0.07503069192171097, -0.11481974273920059, 0.04335082694888115, -0.1704932004213333, -0.042439818382263184, 0.005444696638733149, 0.0139949731528759, 0.013206101022660732, 0.12720820307731628, -0.19255615770816803, -0.01654396951198578, 0.13260798156261444, -0.09212633967399597, -0.118110790848732, 0.07884611934423447, -0.029701577499508858, 0.1624738723039627, 0.04682036489248276, -0.027025915682315826, 0.09224298596382141, -0.16434773802757263, -0.07092688232660294, -0.00949116237461567, -0.01727987825870514, 0.12109188735485077, 0.07512219995260239, -0.05991523340344429, 0.046571120619773865, 0.02832140028476715, -0.038078423589468, -0.04424772411584854, -0.050857074558734894, -0.10884185880422592, -0.01070026308298111, -0.08987759798765182, 0.04065500199794769, -0.01250192429870367, -0.07916021347045898, -0.029885273426771164, -0.18612512946128845, -0.0030564051121473312, 0.10038342326879501, 0.0035033065360039473, -0.005652366206049919, -0.08666291832923889, 0.026358824223279953, -0.03112892620265484, -0.008404186926782131, -0.16764774918556213, -0.04399421438574791, 0.046902090311050415, -0.16094985604286194, 0.020117372274398804, -0.06413903087377548, 0.06334125250577927, 0.03641495108604431, -0.05590536445379257, -0.0248766727745533, -0.01730942726135254, 0.011945613659918308, -0.05083848536014557, -0.18994836509227753, -0.056277405470609665, -0.037882111966609955, 0.149809330701828, -0.25956398248672485, 0.032966937869787216, 0.051140617579221725, 0.14649195969104767, 0.00406361510977149, -0.05115427449345589, 0.01429014839231968, -0.05360214412212372, -0.054652128368616104, -0.06746816635131836, -0.006135428790003061, -0.027576493099331856, -0.05147203803062439, 0.019243421033024788, -0.1755700707435608, -0.021410830318927765, 0.09424154460430145, 0.12876708805561066, -0.1486445665359497, -0.018640631809830666, -0.048725154250860214, -0.06339836865663528, -0.0715010017156601, -0.07038594037294388, 0.10712739825248718, 0.0513901449739933, 0.04796046018600464, -0.07435787469148636, -0.07092321664094925, 0.02726263552904129, 0.006906150374561548, -0.03382374346256256, 0.08727246522903442, 0.05199531093239784, -0.09209315478801727, 0.0756213590502739, 0.1092359870672226, 0.07177663594484329, 0.09363535046577454, 0.01574566215276718, -0.11756632477045059, -0.028492970392107964, 0.036266472190618515, 0.02740776725113392, 0.1465986967086792, -0.05952361226081848, 0.04016614332795143, 0.04494241625070572, -0.04170418903231621, 0.022319864481687546, -0.08787637203931808, 0.024075502529740334, 0.025203049182891846, -0.0034381982404738665, 0.06284574419260025, -0.02525499276816845, -0.0050758360885083675, 0.07016654312610626, 0.047779910266399384, 0.04621000960469246, 0.009655474685132504, -0.01720241829752922, -0.1047825813293457, 0.16950392723083496, -0.0951867327094078, -0.269941508769989, -0.17632324993610382, 0.026197833940386772, 0.04035249724984169, -0.022378476336598396, 0.031619444489479065, -0.07056326419115067, -0.10630585998296738, -0.1060405746102333, -0.002429972169920802, 0.01714223250746727, -0.06364088505506516, -0.0741225928068161, 0.07348573952913284, 0.04382912442088127, -0.14902326464653015, 0.038552410900592804, 0.055694397538900375, -0.057955220341682434, -0.0233661737293005, 0.09118817001581192, 0.12397737801074982, 0.14583967626094818, -0.021366750821471214, -0.028626007959246635, 0.029004426673054695, 0.19620531797409058, -0.13469526171684265, 0.10371150821447372, 0.13814030587673187, -0.04545360431075096, 0.08360563963651657, 0.1560150384902954, 0.029186224564909935, -0.08317049592733383, 0.05044832453131676, 0.04082648828625679, -0.043159641325473785, -0.2666129767894745, -0.0534592866897583, 0.012832709588110447, -0.06255637854337692, 0.09786593168973923, 0.10183793306350708, 0.11542957276105881, 0.034910861402750015, -0.07166364789009094, -0.043925940990448, -0.0058974819257855415, 0.11737963557243347, -0.05490213260054588, -0.012639665976166725, 0.07686592638492584, -0.05086168646812439, 0.005355054512619972, 0.10266812145709991, 0.02973790094256401, 0.17442677915096283, 0.020399179309606552, 0.11231429129838943, 0.06195578724145889, 0.08633565157651901, 0.0007386076031252742, 0.02951662428677082, 0.05147615820169449, 0.017203815281391144, -0.002300140680745244, -0.10421168059110641, -0.006156572140753269, 0.1449710875749588, 0.028103826567530632, 0.029669636860489845, -0.0018948549404740334, -0.005003341939300299, 0.05121048167347908, 0.1746254414319992, -0.011592294089496136, -0.22072425484657288, -0.0845772922039032, 0.06936841458082199, -0.06218599155545235, -0.12968985736370087, -0.026130788028240204, 0.045467354357242584, -0.17519839107990265, 0.026703642681241035, -0.027433741837739944, 0.0919293761253357, -0.09345759451389313, -0.02221956104040146, 0.03687324374914169, 0.084866963326931, -0.014529162086546421, 0.08703910559415817, -0.14498743414878845, 0.11886418610811234, 0.02978132851421833, 0.09024628251791, -0.11081171780824661, 0.07909037172794342, -0.007550720125436783, 0.009180475026369095, 0.19379350543022156, -0.011335089802742004, -0.03514958545565605, -0.08774717897176743, -0.11210042238235474, -0.013537433929741383, 0.12687496840953827, -0.1243172138929367, 0.08773399889469147, -0.015198243781924248, -0.044079482555389404, 0.00937260314822197, -0.12100647389888763, -0.17273177206516266, -0.19628387689590454, 0.05585884302854538, -0.09575839340686798, 0.025643249973654747, -0.11914430558681488, -0.07089093327522278, -0.02952558360993862, 0.241120383143425, -0.1745356321334839, -0.06510113179683685, -0.1468164622783661, -0.046294767409563065, 0.1662203073501587, -0.04437198117375374, 0.0718095526099205, -0.0208172257989645, 0.20345525443553925, 0.005988610442727804, -0.004939318168908358, 0.06724198162555695, -0.08892562240362167, -0.16873881220817566, -0.06771010160446167, 0.1510489284992218, 0.11680185794830322, 0.04907919466495514, -0.002248800592496991, 0.0011772146681323647, -0.016943959519267082, -0.1137804463505745, -0.0033210667315870523, 0.16037839651107788, 0.03878779336810112, 0.025986969470977783, -0.05243593826889992, -0.08797456324100494, -0.06899320334196091, -0.06853509694337845, 0.06221301481127739, 0.19590823352336884, -0.10376439243555069, 0.1700313836336136, 0.147536963224411, -0.07305635511875153, -0.23175598680973053, 0.035342130810022354, 0.04983805492520332, 0.0014306638622656465, 0.04886869341135025, -0.18252557516098022, 0.10521943867206573, 0.019543392583727837, -0.05505957826972008, 0.13485197722911835, -0.1557481735944748, -0.1552847921848297, 0.0722852572798729, 0.03904085233807564, -0.22423844039440155, -0.1354004591703415, -0.09622503817081451, -0.05825018882751465, -0.14065024256706238, 0.06054598465561867, -0.002136280992999673, 0.015948504209518433, 0.03500790148973465, -0.0015643214574083686, 0.027123261243104935, -0.058935679495334625, 0.18609118461608887, -0.004065449349582195, 0.020676052197813988, -0.060264769941568375, -0.0478842556476593, 0.09839435666799545, -0.06130504235625267, 0.12208222597837448, 0.004057085141539574, 0.01594383642077446, -0.10362856835126877, -0.048314861953258514, -0.04328322783112526, 0.05154227837920189, -0.07548051327466965, -0.10070807486772537, -0.043625857681035995, 0.08841723203659058, 0.07005169242620468, -0.03383097052574158, 0.00549331633374095, -0.07189501076936722, 0.10019614547491074, 0.17795267701148987, 0.17573626339435577, 0.009926567785441875, -0.07241068035364151, 0.01677953451871872, -0.04142116755247116, 0.044231921434402466, -0.2513144314289093, 0.03756171092391014, 0.06098250672221184, 0.029438555240631104, 0.09217222779989243, -0.020435843616724014, -0.1820858269929886, -0.04050002992153168, 0.08094815909862518, -0.05452597141265869, -0.22617179155349731, -0.019085140898823738, 0.0954197570681572, -0.2020406424999237, -0.007372708059847355, 0.03995226323604584, -0.048725228756666183, -0.023169852793216705, 0.00010950004070764408, 0.06317184865474701, 0.002471912419423461, 0.09773622453212738, 0.0735151618719101, 0.09715340286493301, -0.08337292820215225, 0.10562895983457565, 0.10150538384914398, -0.09572599828243256, 0.03605884686112404, 0.06754924356937408, -0.05300498008728027, -0.043293699622154236, 0.03665391728281975, 0.033023297786712646, 0.005234600510448217, -0.060321882367134094, 0.013913018628954887, -0.036497246474027634, 0.044923391193151474, 0.08326134830713272, 0.03754979372024536, -0.013354414142668247, 0.06462216377258301, 0.03401726484298706, -0.10898099094629288, 0.10366570204496384, 0.01731540448963642, 0.04105307161808014, -0.08384523540735245, -0.019968897104263306, 0.035425446927547455, 0.030576206743717194, -0.01765924133360386, -0.02306121215224266, -0.02860277332365513, -0.01614218018949032, -0.14299540221691132, -0.023106401786208153, -0.07243485748767853, 0.006181265693157911, 0.014656842686235905, -0.031884219497442245, -0.011233693920075893, 0.02475680410861969, -0.06979699432849884, -0.07426341623067856, -0.006949664559215307, 0.09833318740129471, -0.15115703642368317, 0.008848577737808228, 0.06907843053340912, -0.11088496446609497, 0.08190931379795074, -0.008411259390413761, 0.016245156526565552, 0.022527478635311127, -0.15448406338691711, 0.05601610988378525, 0.0008648968650959432, 0.01916889287531376, 0.025886621326208115, -0.16471809148788452, 0.004104440100491047, -0.04661374166607857, -0.02149827405810356, -0.00004464812809601426, -0.02647159807384014, -0.12325995415449142, 0.06858719140291214, -0.015622655861079693, -0.035931166261434555, -0.02701525390148163, 0.0539589487016201, 0.07888586074113846, -0.027474910020828247, 0.10445091128349304, -0.008690856397151947, 0.04941811040043831, -0.16801609098911285, -0.02470702864229679, -0.04982255399227142, 0.019377702847123146, 0.009884213097393513, -0.007693959400057793, 0.04183054715394974, -0.00976533442735672, 0.21883612871170044, -0.05075952783226967, 0.1607085019350052, 0.05847611650824547, -0.017352959141135216, -0.0007513365126214921, 0.06180921941995621, 0.05997028574347496, 0.04658793285489082, 0.009480604901909828, 0.023740366101264954, -0.022450892254710197, -0.006695089396089315, -0.15932634472846985, 0.01890849508345127, 0.14999441802501678, 0.06301083415746689, 0.024745315313339233, 0.05866100639104843, -0.12775006890296936, -0.12135478109121323, 0.09311001747846603, -0.026755332946777344, 0.00928465835750103, -0.08245618641376495, 0.1358020007610321, 0.14980104565620422, -0.14000412821769714, 0.05256148427724838, -0.06134212389588356, -0.05217423290014267, -0.10388828068971634, -0.12032219022512436, -0.05887215584516525, -0.053666237741708755, 0.002330566756427288, -0.03760887682437897, 0.054546963423490524, 0.03344334661960602, -0.009351172484457493, -0.00022941511997487396, 0.13597318530082703, -0.019751882180571556, -0.0028988157864660025, 0.048313532024621964, 0.03693558648228645, 0.02373051457107067, -0.05275435373187065, 0.02940409444272518, 0.02539868652820587, 0.032232340425252914, 0.06546790152788162, 0.033412106335163116, -0.047448933124542236, 0.03804153576493263, -0.0025254099164158106, -0.11207924783229828, 0.019641218706965446, -0.00460948096588254, -0.0742158442735672, 0.1268945336341858, 0.0407399944961071, 0.010224059224128723, -0.03741471841931343, 0.24361543357372284, -0.06653323769569397, -0.06378097087144852, -0.13251738250255585, 0.10491154342889786, -0.0027236645109951496, 0.06476365029811859, 0.023412218317389488, -0.1284150779247284, 0.005243356805294752, 0.13858191668987274, 0.12181595712900162, 0.0045748427510261536, 0.009228081442415714, 0.0518609918653965, 0.0025186820421367884, -0.06998204439878464, 0.054019294679164886, 0.06992026418447495, 0.12919506430625916, -0.07847554981708527, 0.07680778950452805, 0.0006860480643808842, -0.08370215445756912, -0.02947772853076458, 0.11312682181596756, -0.0409729965031147, 0.03491825982928276, -0.047444481402635574, 0.10916327685117722, -0.05787910893559456, -0.29412412643432617, 0.02350960113108158, -0.09588567912578583, -0.15202060341835022, -0.018367812037467957, 0.05944539234042168, -0.02624768204987049, 0.018029648810625076, 0.06971040368080139, -0.06011629104614258, 0.20098382234573364, 0.0335683599114418, -0.07864278554916382, -0.0664360448718071, 0.04837050288915634, -0.06564252078533173, 0.2949807047843933, 0.008418165147304535, 0.02863333560526371, 0.10770907253026962, -0.03253700211644173, -0.18271861970424652, 0.010723991319537163, 0.1133992001414299, -0.08056149631738663, 0.08200647681951523, 0.19000613689422607, -0.012578671798110008, 0.1209007054567337, 0.05294662341475487, -0.047376248985528946, 0.04217283055186272, -0.03389401361346245, -0.051268599927425385, -0.10752558708190918, 0.058453381061553955, -0.05909625440835953, 0.15447644889354706, 0.10152646154165268, -0.05671518296003342, -0.004550917539745569, -0.05555408447980881, 0.04875178262591362, 0.01804669201374054, 0.12263146042823792, 0.02951994352042675, -0.1865430772304535, 0.032826557755470276, -0.01144319772720337, 0.10186848044395447, -0.25588861107826233, -0.08421015739440918, 0.08833149075508118, -0.011924264021217823, -0.05105875805020332, 0.10560628771781921, 0.057650718837976456, 0.04243382066488266, -0.043439045548439026, -0.10480839014053345, -0.02186836116015911, 0.14663739502429962, -0.1469624787569046, -0.025013303384184837 ]
null
null
null
# **Q-Learning** Agent playing1 **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="aturja65/q-Taxi-v3", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
{"tags": ["Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "q-Taxi-v3", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "Taxi-v3", "type": "Taxi-v3"}, "metrics": [{"type": "mean_reward", "value": "7.56 +/- 2.71", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
aljaziz/q-Taxi-v3
[ "Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
2024-02-10T13:05:05+00:00
[]
[]
TAGS #Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
# Q-Learning Agent playing1 Taxi-v3 This is a trained model of a Q-Learning agent playing Taxi-v3 . ## Usage
[ "# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage" ]
[ "TAGS\n#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n", "# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage" ]
[ 32, 33 ]
[ "passage: TAGS\n#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage" ]
[ 0.048862796276807785, -0.16549694538116455, -0.005485367961227894, 0.02960980497300625, 0.1345081776380539, -0.01784728653728962, 0.11895976960659027, 0.07759871333837509, -0.07461097836494446, -0.055395450443029404, 0.1418241262435913, 0.09088201075792313, 0.055222880095243454, 0.05699880048632622, 0.09511256217956543, -0.27440664172172546, 0.048217080533504486, -0.02918700873851776, 0.05621987581253052, 0.11878681182861328, 0.0670095682144165, -0.040441032499074936, 0.061956584453582764, 0.11818158626556396, -0.1018151044845581, -0.007344264071434736, 0.035402704030275345, -0.09440053254365921, 0.17413531243801117, 0.07204403728246689, 0.12337774783372879, 0.05132639780640602, 0.179361954331398, -0.12762396037578583, 0.024310702458024025, -0.0010275895474478602, -0.10138072073459625, -0.03909514099359512, -0.012415820732712746, -0.08349097520112991, 0.03230205550789833, 0.23522862792015076, 0.07199250161647797, 0.06632792949676514, -0.17707863450050354, -0.06584878265857697, -0.04375573247671127, 0.069611094892025, 0.14951466023921967, 0.03758616745471954, -0.033800311386585236, 0.1684885323047638, -0.2564343810081482, 0.05066783353686333, 0.037275806069374084, -0.42313119769096375, 0.017119819298386574, 0.1507398933172226, 0.15090937912464142, 0.06909667700529099, -0.10573802888393402, 0.013512322679162025, 0.051325585693120956, -0.0005318621988408267, 0.024325110018253326, 0.006554204970598221, 0.15601307153701782, 0.08537693321704865, -0.1487821787595749, -0.058576688170433044, 0.17441977560520172, -0.03788546845316887, -0.02613203600049019, -0.039745692163705826, 0.0067160045728087425, -0.06427708268165588, -0.004067842848598957, -0.1777995079755783, 0.00734262028709054, 0.06666424125432968, -0.014348524622619152, 0.014901017770171165, -0.035522811114788055, -0.0966939702630043, -0.023098144680261612, -0.08592145889997482, 0.01677769608795643, -0.006319406442344189, -0.10187895596027374, 0.05002119392156601, -0.061138734221458435, 0.0014382408699020743, -0.05123179033398628, -0.15047866106033325, -0.049055423587560654, -0.03481535613536835, 0.1474713832139969, -0.0044205985032022, -0.01873963139951229, -0.03164304047822952, 0.15474793314933777, 0.049551334232091904, -0.05370146036148071, 0.05625450983643532, 0.07605006545782089, 0.23867930471897125, 0.10401605814695358, 0.10196955502033234, -0.06798075139522552, 0.10180158913135529, -0.12330973148345947, -0.08915644884109497, -0.17508824169635773, 0.11820860952138901, 0.00015364694991149008, 0.1317785084247589, -0.12023144960403442, 0.07898581773042679, -0.067511186003685, 0.013453764840960503, 0.01636839471757412, 0.0820009782910347, -0.012399360537528992, 0.10676060616970062, -0.005061192903667688, -0.06941985338926315, 0.014177112840116024, 0.05935845896601677, 0.03754841163754463, -0.038601722568273544, -0.03192409873008728, -0.05762290954589844, -0.05065649375319481, -0.10128600150346756, -0.06447898596525192, 0.018573462963104248, -0.007677143905311823, -0.1833900660276413, -0.06407523155212402, 0.00897200871258974, 0.015712225809693336, -0.03988850116729736, -0.05148044601082802, -0.15265507996082306, -0.042461175471544266, -0.015450406819581985, -0.03500641882419586, -0.06214277446269989, -0.0383245050907135, 0.046435944736003876, -0.07560601085424423, 0.013364278711378574, 0.023342855274677277, 0.05405820533633232, -0.025881100445985794, 0.06068144738674164, -0.08357544988393784, 0.09493788331747055, -0.1540430635213852, -0.03271956741809845, -0.025445878505706787, -0.041183918714523315, 0.1752462536096573, 0.06099751964211464, -0.015994304791092873, 0.15260063111782074, -0.17141541838645935, -0.058121129870414734, 0.15596486628055573, 0.008629098534584045, -0.09967197477817535, -0.003560945624485612, -0.09397093951702118, 0.1428760588169098, 0.08571921288967133, 0.2478504776954651, 0.12005335837602615, -0.22748184204101562, 0.055358242243528366, 0.12515293061733246, -0.14365963637828827, 0.10365243256092072, 0.07344598323106766, 0.005470725707709789, -0.18886831402778625, -0.06843198090791702, -0.06121627986431122, 0.1053021252155304, -0.08522345870733261, -0.0776243582367897, 0.09323626756668091, -0.05086790770292282, 0.24641476571559906, -0.028281206265091896, 0.06174173951148987, -0.026681531220674515, -0.1389324963092804, -0.01723906397819519, 0.060955192893743515, 0.05258452147245407, -0.024835573509335518, -0.25895482301712036, 0.13646544516086578, 0.048650871962308884, 0.025074828416109085, 0.004106190986931324, -0.05691491439938545, 0.016934165731072426, 0.1511998474597931, 0.020012924447655678, 0.13717477023601532, 0.027723990380764008, 0.0706823319196701, -0.006239562761038542, -0.10560829937458038, -0.04169593006372452, 0.061916545033454895, -0.08518962562084198, -0.06641357392072678, 0.011197872459888458, -0.06935211271047592, -0.11783787608146667, -0.12166737765073776, -0.026334572583436966, -0.02980303019285202, -0.07444227486848831, 0.02368103712797165, 0.06536602973937988, -0.06702698022127151, -0.0023908785078674555, 0.007125476840883493, -0.011537045240402222, 0.16434046626091003, 0.011393417604267597, -0.007796820718795061, 0.1328643560409546, -0.11533161997795105, 0.12461213022470474, 0.049438029527664185, -0.024806302040815353, -0.04662557691335678, 0.0014137453399598598, -0.057529181241989136, 0.029044216498732567, -0.04390640929341316, 0.02774495631456375, 0.20111067593097687, 0.02772962674498558, 0.11389166116714478, -0.0656520202755928, 0.04385066404938698, -0.007961965166032314, -0.009693224914371967, 0.018563594669103622, 0.07608018070459366, 0.07813210040330887, -0.1324140727519989, 0.02262016013264656, 0.22455167770385742, 0.1385764330625534, 0.18313980102539062, -0.010877152904868126, 0.06325667351484299, -0.04875868931412697, 0.027505528181791306, 0.024100203067064285, 0.10314226150512695, -0.10732068121433258, -0.0322517491877079, -0.025407759472727776, 0.023599207401275635, -0.08197105675935745, -0.1055799350142479, -0.090115025639534, 0.01222382951527834, -0.03125503659248352, -0.15570329129695892, 0.13300658762454987, -0.10451057553291321, 0.01802753657102585, 0.04692702740430832, -0.22163605690002441, 0.11530312895774841, 0.014291439205408096, -0.10303618758916855, 0.11281087249517441, -0.12051989883184433, -0.08699832111597061, -0.05777236074209213, -0.18658851087093353, 0.05280197039246559, 0.04673841595649719, 0.05166793242096901, -0.18521739542484283, 0.024835903197526932, 0.05545609071850777, 0.13426995277404785, -0.09743253141641617, -0.07142634689807892, -0.15038461983203888, 0.016068490222096443, -0.033661190420389175, -0.16029728949069977, -0.005609163548797369, -0.032781440764665604, -0.18849676847457886, -0.04539939761161804, -0.15086813271045685, -0.034627582877874374, 0.20464378595352173, 0.026907702907919884, 0.09480511397123337, -0.07926445454359055, 0.3802889585494995, -0.042039383202791214, -0.06146497279405594, -0.01321389526128769, -0.07072482258081436, 0.02512686513364315, 0.13271741569042206, 0.0036099457647651434, -0.017886579036712646, -0.0037857077550143003, 0.0024592927657067776, -0.06234965845942497, -0.13400450348854065, 0.0028710351325571537, 0.03905198723077774, 0.1874423623085022, 0.004639793653041124, 0.06659388542175293, 0.03133883699774742, 0.057546284049749374, 0.07748064398765564, 0.030926106497645378, 0.0011591583024710417, -0.01591806672513485, 0.06604493409395218, -0.11684755235910416, 0.042466625571250916, -0.030429253354668617, -0.10143838077783585, -0.013183288276195526, 0.07950251549482346, 0.12755028903484344, 0.17849206924438477, -0.04790908098220825, 0.17489230632781982, 0.13580141961574554, 0.16576050221920013, 0.049315933138132095, -0.020801831036806107, -0.08773037046194077, -0.06118565797805786, 0.004774159751832485, -0.031952597200870514, 0.04869702458381653, 0.3231290578842163, 0.037619613111019135, -0.09036035090684891, 0.11149907857179642, 0.009480619803071022, 0.05359881371259689, 0.022797370329499245, -0.11162138730287552, 0.11170321702957153, 0.07968773692846298, -0.06341761350631714, -0.07602835446596146, 0.16758501529693604, -0.1109386757016182, -0.26646625995635986, -0.11410990357398987, -0.012305386364459991, 0.07903840392827988, 0.005651174578815699, 0.05498376116156578, -0.11829282343387604, -0.16034497320652008, -0.034191906452178955, 0.1335442066192627, -0.3077351450920105, 0.2065143585205078, -0.0198091771453619, 0.06707923114299774, -0.039657969027757645, -0.07026876509189606, 0.09694647043943405, 0.13174086809158325, 0.29124146699905396, 0.01396956667304039, 0.04841272905468941, -0.15176129341125488, -0.0976925864815712, 0.0018439020495861769, 0.015482662245631218, -0.02563396655023098, 0.028520405292510986, -0.0540912002325058, 0.008404579944908619, -0.018086453899741173, 0.2102297693490982, -0.11316607892513275, 0.004344627261161804, -0.06968966871500015, -0.11707738786935806, 0.19409789144992828, -0.07178345322608948, -0.04543264955282211, -0.14959357678890228, -0.15512511134147644, -0.004174166824668646, -0.02413962036371231, -0.019664527848362923, -0.17603960633277893, -0.18804074823856354, -0.05204557999968529, -0.005645004566758871, -0.003464865731075406, 0.05867868289351463, -0.07517234236001968, -0.04805335775017738, 0.1009904220700264, -0.07743175327777863, -0.056063808500766754, -0.1103200614452362, 0.1391381323337555, 0.06248528137803078, 0.16743235290050507, 0.05907081440091133, 0.0006117874872870743, 0.11471151560544968, -0.02913086675107479, 0.11103474348783493, -0.11291708797216415, -0.17145049571990967, -0.08334989100694656, -0.018775060772895813, 0.09519003331661224, -0.04789286106824875, 0.0028788831550627947, 0.2550160884857178, 0.14880181849002838, -0.0897710770368576, 0.27680760622024536, 0.04414956644177437, -0.09375058114528656, -0.18432219326496124, -0.15961645543575287, 0.03759992495179176, 0.060025621205568314, 0.13095876574516296, -0.057205069810152054, -0.08483537286520004, -0.08492398262023926, -0.07478608191013336, -0.13140805065631866, -0.24232175946235657, -0.030598774552345276, 0.22874866425991058, 0.08656918257474899, 0.08219650387763977, -0.012482990510761738, -0.01186054851859808, 0.00526038184762001, 0.02680150233209133, 0.12018456310033798, -0.13341329991817474, 0.11107480525970459, 0.022198403254151344, 0.044267985969781876, 0.009712530300021172, 0.07929777354001999, 0.03375575691461563, -0.003218587953597307, -0.0006439819699153304, -0.0988350659608841, -0.2596651017665863, 0.0816885456442833, -0.01623627357184887, -0.09960969537496567, 0.014988959766924381, 0.02061903104186058, -0.2089255303144455, 0.011128270998597145, -0.019883770495653152, -0.03150356933474541, -0.06483490765094757, -0.10664787143468857, -0.056551624089479446, 0.04928823933005333, 0.10853826254606247, 0.011660109274089336, 0.05354316532611847, -0.0404130220413208, 0.07917837053537369, 0.0826287642121315, 0.15132710337638855, 0.06795957684516907, -0.190711110830307, -0.10953907668590546, -0.0414445661008358, 0.12121522426605225, -0.12505418062210083, 0.036917757242918015, 0.053161121904850006, -0.016534561291337013, 0.14621229469776154, 0.1070784479379654, -0.07452095299959183, 0.11915595084428787, 0.08904775977134705, -0.04094788804650307, -0.23367151618003845, -0.07120766490697861, 0.11133213341236115, 0.07195597887039185, -0.03961895406246185, 0.018120890483260155, -0.04960581287741661, -0.013980977237224579, 0.048759616911411285, -0.0538676381111145, -0.07230538129806519, 0.004421027842909098, 0.1247575581073761, 0.1029362753033638, -0.04655474051833153, 0.01296416949480772, 0.037371400743722916, 0.003788623260334134, 0.04730486497282982, 0.0407949760556221, -0.08269952982664108, -0.04124005511403084, 0.02782733179628849, 0.37552911043167114, -0.010165480896830559, -0.020456433296203613, 0.018555615097284317, -0.19949445128440857, 0.09135842323303223, 0.13205479085445404, 0.04697350412607193, 0.004247748292982578, -0.08139242231845856, 0.026877427473664284, -0.010625290684401989, 0.09936143457889557, -0.07806670665740967, -0.05493134260177612, -0.21631066501140594, -0.025010565295815468, 0.017490221187472343, 0.24077683687210083, -0.08458559215068817, -0.12801732122898102, -0.20628872513771057, 0.13128381967544556, -0.11333390325307846, -0.03695881739258766, -0.024473199620842934, 0.03926658630371094, -0.01989821158349514, 0.06291737407445908, -0.0710630789399147, 0.006373001262545586, -0.11024709790945053, 0.055267609655857086, 0.04204455390572548, 0.1229788213968277, 0.014207782223820686, 0.02016810141503811, 0.05822525918483734, -0.01837925612926483, 0.07173580676317215, -0.06203491613268852, -0.04550490900874138, 0.14224006235599518, -0.020255116745829582, -0.04152837023139, -0.0483345128595829, -0.036874305456876755, 0.11981741338968277, -0.05059147998690605, -0.007141099311411381, -0.054929375648498535, -0.06906463205814362, 0.03462086617946625, -0.009175732731819153, -0.008798843249678612, 0.06801853328943253, 0.04024988040328026, -0.026994358748197556, 0.005263668950647116, 0.03447828069329262, -0.10330043733119965, -0.04955084249377251, 0.16955432295799255, -0.0749620869755745, 0.10274054110050201, -0.031069839373230934, 0.018015999346971512, 0.005847334861755371, -0.022399673238396645, -0.015360680408775806, -0.1457086056470871, -0.06137600541114807, -0.09489979594945908, 0.11565322428941727, 0.08146517723798752, 0.03358805552124977, 0.04274565726518631, 0.019532648846507072, -0.04414922371506691, -0.038583990186452866, 0.12961317598819733, 0.08133101463317871, 0.012996876612305641, 0.01137041300535202, 0.01941833831369877, -0.020302120596170425, 0.0028480992186814547, -0.01250747125595808, -0.07239153981208801, -0.05874783173203468, 0.09400010108947754, 0.1600283533334732, -0.06127211079001427, -0.13325586915016174, -0.020593497902154922, 0.04988488554954529, 0.0014717020094394684, -0.08777432143688202, 0.04833676666021347, 0.15805292129516602, -0.05623878911137581, 0.03216489031910896, -0.09984751045703888, -0.07263360917568207, -0.16060975193977356, -0.10029061883687973, -0.06092562898993492, -0.28350353240966797, 0.09752398729324341, 0.006392303854227066, -0.014731393195688725, 0.059529416263103485, 0.051305368542671204, -0.052508849650621414, 0.07068239152431488, -0.18146829307079315, -0.007054794579744339, 0.03497592359781265, -0.13212306797504425, 0.02475893869996071, -0.2378365397453308, 0.10198072344064713, -0.04623803123831749, -0.1519704908132553, -0.04004510119557381, 0.0641569048166275, -0.09540136158466339, -0.01822364516556263, -0.0475153923034668, -0.01922670193016529, 0.01624443754553795, -0.009348669089376926, -0.031147832050919533, 0.13716529309749603, 0.02827494591474533, -0.03268734738230705, 0.005254602525383234, 0.0223685409873724, 0.03955082967877388, -0.0969657450914383, -0.05986930429935455, 0.08311155438423157, -0.031056145206093788, 0.14728976786136627, 0.000341245875461027, 0.04181376099586487, -0.06758682429790497, 0.2593761384487152, 0.2023983597755432, -0.12479214370250702, 0.008118697442114353, -0.021801479160785675, 0.012670028023421764, -0.041751839220523834, 0.13110700249671936, 0.013386172242462635, 0.12186761200428009, -0.17513342201709747, -0.01036517322063446, -0.0818324014544487, -0.04501292482018471, 0.06702108681201935, 0.14714950323104858, 0.15742522478103638, 0.03436789661645889, -0.07328428328037262, 0.06722653657197952, -0.30119743943214417, 0.20540550351142883, -0.1346001923084259, -0.01498429011553526, -0.040251150727272034, -0.058389630168676376, 0.061147745698690414, 0.11309876292943954, 0.10832664370536804, -0.021150551736354828, -0.0905047357082367, -0.04486766457557678, -0.039378076791763306, -0.13019338250160217, -0.02718670479953289, 0.1654091775417328, 0.06799814850091934, 0.31520840525627136, -0.017577875405550003, 0.07702425122261047, 0.034410297870635986, 0.06451138854026794, 0.004519328009337187, 0.09537279605865479, 0.07960964739322662, -0.06345855444669724, -0.07373003661632538, -0.001637450186535716, 0.05033271387219429, 0.14567798376083374, -0.03826142102479935, -0.18691548705101013, 0.15858715772628784, 0.07192251086235046, -0.13762691617012024, -0.05777517706155777, 0.08409425616264343, -0.0739973932504654, 0.0550808347761631, 0.08115427941083908, 0.015876613557338715, -0.017793258652091026, -0.004664506763219833, 0.06074233725667, 0.024694660678505898, -0.02343848906457424, 0.003570882137864828, -0.08337053656578064, -0.04151543974876404, 0.07267895340919495, -0.0844460055232048, -0.20546193420886993, -0.0957019031047821, -0.07551700621843338, 0.030557552352547646, -0.0649830624461174, 0.12575586140155792, 0.1717868149280548, 0.0593598335981369, -0.03307248651981354, -0.10721943527460098, -0.035562749952077866, 0.07602505385875702, -0.044773899018764496, -0.09409699589014053 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Qwen1.5-1.8B-OpenHermes-2.5 This model is a fine-tuned version of [Qwen/Qwen1.5-1.8B](https://huggingface.co/Qwen/Qwen1.5-1.8B) on the teknium/OpenHermes-2.5 dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - distributed_type: multi-GPU - num_devices: 2 - gradient_accumulation_steps: 16 - total_train_batch_size: 128 - total_eval_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 1.0 ### Training results ### Framework versions - Transformers 4.37.2 - Pytorch 2.0.1+cu117 - Datasets 2.16.1 - Tokenizers 0.15.1 ### Inference ``` from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline model_id = "minghaowu/phi-2-OpenHermes-2.5" tokenizer = AutoTokenizer.from_pretrained(model_id) model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.bfloat16, device_map="auto") pipe = pipeline("text-generation", model=model, tokenizer=tokenizer, device_map="auto") your_instruction = <your_instruction> infer_prompt = f"### USER: {your_instruction} <|endoftext|>\n### ASSISTANT:" output = pipe(infer_prompt, do_sample=True, max_new_tokens=256)[0]["generated_text"] print(output) ```
{"license": "other", "tags": ["generated_from_trainer"], "datasets": ["teknium/OpenHermes-2.5"], "base_model": "Qwen/Qwen1.5-1.8B", "model-index": [{"name": "Qwen1.5-1.8B-OpenHermes-2.5", "results": []}]}
text-generation
minghaowu/Qwen1.5-1.8B-OpenHermes-2.5
[ "transformers", "safetensors", "qwen2", "text-generation", "generated_from_trainer", "conversational", "dataset:teknium/OpenHermes-2.5", "base_model:Qwen/Qwen1.5-1.8B", "license:other", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-10T13:06:34+00:00
[]
[]
TAGS #transformers #safetensors #qwen2 #text-generation #generated_from_trainer #conversational #dataset-teknium/OpenHermes-2.5 #base_model-Qwen/Qwen1.5-1.8B #license-other #autotrain_compatible #endpoints_compatible #region-us
# Qwen1.5-1.8B-OpenHermes-2.5 This model is a fine-tuned version of Qwen/Qwen1.5-1.8B on the teknium/OpenHermes-2.5 dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 4 - eval_batch_size: 8 - seed: 42 - distributed_type: multi-GPU - num_devices: 2 - gradient_accumulation_steps: 16 - total_train_batch_size: 128 - total_eval_batch_size: 16 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 1.0 ### Training results ### Framework versions - Transformers 4.37.2 - Pytorch 2.0.1+cu117 - Datasets 2.16.1 - Tokenizers 0.15.1 ### Inference
[ "# Qwen1.5-1.8B-OpenHermes-2.5\n\nThis model is a fine-tuned version of Qwen/Qwen1.5-1.8B on the teknium/OpenHermes-2.5 dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 8\n- seed: 42\n- distributed_type: multi-GPU\n- num_devices: 2\n- gradient_accumulation_steps: 16\n- total_train_batch_size: 128\n- total_eval_batch_size: 16\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_ratio: 0.1\n- num_epochs: 1.0", "### Training results", "### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.0.1+cu117\n- Datasets 2.16.1\n- Tokenizers 0.15.1", "### Inference" ]
[ "TAGS\n#transformers #safetensors #qwen2 #text-generation #generated_from_trainer #conversational #dataset-teknium/OpenHermes-2.5 #base_model-Qwen/Qwen1.5-1.8B #license-other #autotrain_compatible #endpoints_compatible #region-us \n", "# Qwen1.5-1.8B-OpenHermes-2.5\n\nThis model is a fine-tuned version of Qwen/Qwen1.5-1.8B on the teknium/OpenHermes-2.5 dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 8\n- seed: 42\n- distributed_type: multi-GPU\n- num_devices: 2\n- gradient_accumulation_steps: 16\n- total_train_batch_size: 128\n- total_eval_batch_size: 16\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_ratio: 0.1\n- num_epochs: 1.0", "### Training results", "### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.0.1+cu117\n- Datasets 2.16.1\n- Tokenizers 0.15.1", "### Inference" ]
[ 81, 45, 6, 12, 8, 3, 157, 4, 33, 5 ]
[ "passage: TAGS\n#transformers #safetensors #qwen2 #text-generation #generated_from_trainer #conversational #dataset-teknium/OpenHermes-2.5 #base_model-Qwen/Qwen1.5-1.8B #license-other #autotrain_compatible #endpoints_compatible #region-us \n# Qwen1.5-1.8B-OpenHermes-2.5\n\nThis model is a fine-tuned version of Qwen/Qwen1.5-1.8B on the teknium/OpenHermes-2.5 dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 4\n- eval_batch_size: 8\n- seed: 42\n- distributed_type: multi-GPU\n- num_devices: 2\n- gradient_accumulation_steps: 16\n- total_train_batch_size: 128\n- total_eval_batch_size: 16\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- lr_scheduler_warmup_ratio: 0.1\n- num_epochs: 1.0### Training results### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.0.1+cu117\n- Datasets 2.16.1\n- Tokenizers 0.15.1### Inference" ]
[ -0.10150772333145142, 0.18253980576992035, -0.003545155515894294, 0.07475906610488892, 0.09023259580135345, 0.03941012918949127, 0.09200479835271835, 0.1539490520954132, -0.034231435507535934, 0.11628200113773346, 0.061100561171770096, 0.023559581488370895, 0.0849762037396431, 0.1146286278963089, 0.008505710400640965, -0.2199392020702362, 0.020250393077731133, -0.048189129680395126, -0.0409836508333683, 0.06645146757364273, 0.09195773303508759, -0.0722164660692215, 0.05391858145594597, -0.014491809532046318, -0.0812772810459137, -0.01168894674628973, -0.0490613654255867, -0.05256111919879913, 0.0853710025548935, 0.03763999417424202, 0.027420710772275925, 0.01681268773972988, 0.11472971737384796, -0.27892521023750305, -0.0005542043363675475, 0.07881494611501694, 0.010739930905401707, 0.06714504212141037, 0.09090987592935562, 0.010774241760373116, 0.08717382699251175, -0.15503504872322083, 0.08986061066389084, 0.04522779583930969, -0.06626569479703903, -0.1640111207962036, -0.08730323612689972, 0.09843432903289795, 0.09010305255651474, 0.05756872147321701, -0.005395156797021627, 0.12365103513002396, -0.06495018303394318, 0.06255913525819778, 0.16081376373767853, -0.3070615530014038, -0.04708733409643173, 0.0790412649512291, 0.058375149965286255, 0.0575997494161129, -0.12073227763175964, -0.003450091928243637, 0.036674864590168, 0.01626729406416416, 0.061461906880140305, 0.012997973710298538, -0.01447498332709074, 0.007737114559859037, -0.12556391954421997, -0.01796969585120678, 0.13121956586837769, 0.07080698013305664, -0.03370298072695732, -0.14245887100696564, -0.036033980548381805, -0.08444493263959885, -0.008959767408668995, -0.028813233599066734, 0.00783954095095396, -0.041061289608478546, -0.02681056596338749, -0.02743895724415779, -0.04970330372452736, -0.05258350819349289, 0.016207309439778328, 0.09819953888654709, 0.057811763137578964, -0.004984761588275433, 0.029150942340493202, 0.10477271676063538, -0.05659059062600136, -0.1223190426826477, -0.04914834722876549, -0.018159056082367897, -0.1486998200416565, -0.053583186119794846, -0.011798151768743992, 0.029209500178694725, 0.032076526433229446, 0.18067748844623566, -0.05431940406560898, 0.07887842506170273, 0.032904647290706635, -0.01522728893905878, 0.015649059787392616, 0.11019521206617355, -0.06436144560575485, -0.12279978394508362, -0.009365081787109375, 0.09813502430915833, 0.022351983934640884, -0.020922290161252022, -0.07727641612291336, -0.041236136108636856, 0.087247334420681, 0.07571607083082199, 0.014398571103811264, 0.029937241226434708, -0.03968903794884682, -0.035969946533441544, 0.09228668361902237, -0.1259426325559616, 0.05376996845006943, -0.0007770509109832346, -0.053758494555950165, -0.04483375325798988, 0.007010809611529112, 0.006961320526897907, -0.0454571433365345, 0.045015547424554825, -0.06359227001667023, -0.0355234369635582, -0.04673224687576294, -0.04407259449362755, 0.045582354068756104, -0.06651972979307175, -0.009645896963775158, -0.07282944023609161, -0.15762226283550262, -0.032088007777929306, 0.03620627522468567, -0.09188589453697205, -0.057715389877557755, -0.029918517917394638, -0.05843416228890419, 0.028138542547822, -0.003070131642743945, 0.09111353009939194, -0.04216262698173523, 0.05707420036196709, 0.0028464801143854856, 0.03692813590168953, 0.06888286024332047, 0.016351334750652313, -0.047606851905584335, 0.058321405202150345, -0.10154443979263306, 0.0938289612531662, -0.11198239773511887, 0.014320863410830498, -0.13375748693943024, -0.08221535384654999, -0.0016691457713022828, -0.040804799646139145, 0.07139606773853302, 0.1264636069536209, -0.13967029750347137, -0.0020904610864818096, 0.1863539069890976, -0.06695958971977234, -0.11391954869031906, 0.1268882155418396, -0.015799030661582947, -0.025903068482875824, 0.03829647973179817, 0.15891234576702118, 0.1380324810743332, -0.15313228964805603, -0.05656068027019501, 0.03261386975646019, 0.08545944094657898, 0.060316652059555054, 0.09451532363891602, -0.013679854571819305, 0.07202316075563431, 0.017604030668735504, -0.034805797040462494, -0.0018248737324029207, -0.06354403495788574, -0.08842287957668304, -0.05065007507801056, -0.09976959228515625, 0.037475600838661194, 0.013899614103138447, 0.0063039446249604225, -0.057703897356987, -0.13572236895561218, 0.01699957251548767, 0.14836741983890533, -0.04439033195376396, 0.009649230167269707, -0.1019473671913147, 0.049113187938928604, -0.026509320363402367, 0.004156481474637985, -0.15591590106487274, -0.1321098357439041, 0.06317176669836044, -0.11269214749336243, 0.017889274284243584, -0.009136956185102463, 0.05122911557555199, 0.08426208794116974, -0.06107088923454285, -0.027340034022927284, -0.061683036386966705, 0.00041148997843265533, -0.07429881393909454, -0.19852015376091003, -0.07580144703388214, -0.03714212402701378, 0.17742858827114105, -0.2206006646156311, 0.0182279534637928, 0.0058737159706652164, 0.13063721358776093, 0.010851157829165459, -0.06143907830119133, 0.012827702797949314, 0.015764111652970314, -0.020326578989624977, -0.09741820394992828, 0.014504877850413322, 0.0029768964741379023, -0.1057470515370369, -0.0488467775285244, -0.15345561504364014, 0.058359161019325256, 0.05292042717337608, 0.09145520627498627, -0.09018844366073608, -0.07246330380439758, -0.05668782815337181, -0.05751713737845421, -0.05841626971960068, -0.010849334299564362, 0.15770423412322998, 0.016296038404107094, 0.0729813203215599, -0.06666210293769836, -0.08379441499710083, 0.016583742573857307, 0.021189356222748756, -0.02701493725180626, 0.0900563895702362, 0.06370732188224792, -0.1176375299692154, 0.0651116743683815, 0.07586334645748138, -0.006593787577003241, 0.1007959395647049, -0.04048222675919533, -0.07794010639190674, -0.0529436394572258, 0.04405170679092407, 0.011794374324381351, 0.13072119653224945, -0.11473516374826431, 0.012874500826001167, 0.0373687706887722, 0.020146086812019348, 0.02142910100519657, -0.11532769352197647, 0.004105886444449425, 0.022098053246736526, -0.04034111648797989, -0.00906817615032196, -0.03620491176843643, 0.0002100530982716009, 0.08303643018007278, 0.028199579566717148, 0.043805014342069626, 0.0049362932331860065, -0.023907626047730446, -0.09734466671943665, 0.1812620908021927, -0.11565827578306198, -0.1125105544924736, -0.10399781912565231, 0.024006973952054977, -0.05856860801577568, -0.029738884419202805, 0.004143528174608946, -0.08173339813947678, -0.05580711364746094, -0.09455876797437668, -0.007674716878682375, -0.0443095825612545, -0.004020955879241228, 0.055074866861104965, 0.04171319305896759, 0.07544248551130295, -0.11978349834680557, 0.021786879748106003, 0.018031954765319824, -0.08715296536684036, 0.0004637128149624914, 0.061016108840703964, 0.10511990636587143, 0.1189194992184639, 0.02408326044678688, 0.010339407250285149, -0.035648416727781296, 0.20476245880126953, -0.08328799903392792, 0.038489434868097305, 0.07785630226135254, 0.0005013123736716807, 0.04733753204345703, 0.11900655925273895, 0.026184268295764923, -0.08798225969076157, 0.007178907748311758, 0.04991953819990158, -0.01580209843814373, -0.2506313621997833, -0.05574469268321991, -0.029736177995800972, -0.0036347932182252407, 0.09829790890216827, 0.054043326526880264, -0.0000038623143154836725, 0.05056411027908325, -0.04352482408285141, 0.05547083914279938, -0.01052236370742321, 0.08559594303369522, 0.06739875674247742, 0.044184934347867966, 0.07693957537412643, -0.038602545857429504, 0.023073779419064522, 0.07454901188611984, 0.06882009655237198, 0.22476151585578918, -0.04225776344537735, 0.15902625024318695, 0.019542429596185684, 0.17909957468509674, -0.03534352034330368, 0.007975740358233452, 0.018339728936553, -0.0019962789956480265, 0.02280559204518795, -0.07387422025203705, -0.024993225932121277, 0.06142845377326012, 0.003318802686408162, -0.0021238697227090597, -0.06954319030046463, 0.06567087769508362, 0.03759967163205147, 0.18639473617076874, 0.05060417950153351, -0.24327890574932098, -0.09077069163322449, 0.03377355635166168, -0.02727765031158924, -0.05264569818973541, 0.014489452354609966, 0.12599046528339386, -0.12899869680404663, 0.08362838625907898, -0.03337223827838898, 0.07799786329269409, -0.07173249125480652, -0.01936793327331543, 0.053705811500549316, 0.10402487963438034, 0.004598062019795179, 0.10875875502824783, -0.16044536232948303, 0.18154238164424896, 0.027451973408460617, 0.09445556253194809, -0.07891269028186798, 0.05555179715156555, -0.013198180124163628, 0.05397346615791321, 0.12677383422851562, 0.005337574519217014, -0.061695609241724014, -0.14406482875347137, -0.10652727633714676, 0.028411908075213432, 0.1300313025712967, -0.04553813114762306, 0.10172244906425476, -0.029997646808624268, -0.009293487295508385, 0.02502044290304184, -0.03821907564997673, -0.16314153373241425, -0.173177108168602, 0.04839574918150902, -0.024151673540472984, -0.022243190556764603, -0.08162809908390045, -0.06848841905593872, -0.10008007287979126, 0.20874568819999695, -0.07037743180990219, -0.04820932820439339, -0.13355164229869843, 0.0709373950958252, 0.1441250592470169, -0.07819438725709915, 0.0184131171554327, 0.005241733975708485, 0.12514334917068481, 0.005754106678068638, -0.06397592276334763, 0.06285605579614639, -0.08784694969654083, -0.17001140117645264, -0.05586069077253342, 0.13657040894031525, 0.0466582216322422, 0.05726010352373123, 0.010797825641930103, -0.005259237252175808, 0.03364932909607887, -0.11207625269889832, 0.043246760964393616, 0.08812123537063599, 0.05546159669756889, 0.020035861060023308, -0.07288143038749695, 0.01877737231552601, -0.03696053847670555, -0.029639702290296555, 0.11073663085699081, 0.2524509131908417, -0.1013440489768982, 0.09198813140392303, 0.09252964705228806, -0.07285613566637039, -0.1556965410709381, 0.013685216195881367, 0.09939835965633392, 0.007517391815781593, 0.036766450852155685, -0.19184510409832, 0.09777648746967316, 0.09625638276338577, -0.0236272681504488, 0.06357018649578094, -0.2963646352291107, -0.12211671471595764, 0.04461122304201126, 0.06495925784111023, -0.06440982222557068, -0.16892719268798828, -0.07121049612760544, -0.009428310208022594, -0.15969844162464142, 0.07879764586687088, -0.053922343999147415, 0.08502742648124695, 0.004716311115771532, 0.030818747356534004, 0.03671148419380188, -0.052913568913936615, 0.1609622985124588, 0.04950210079550743, 0.023297732695937157, -0.06850062310695648, 0.04875202104449272, 0.08751463890075684, -0.07476412504911423, 0.054256029427051544, -0.03783726692199707, 0.06433990597724915, -0.1467866152524948, -0.029538867995142937, -0.041243478655815125, 0.046260278671979904, -0.06538864970207214, -0.06421279907226562, -0.03512616083025932, 0.07960264384746552, 0.08633089065551758, -0.0283063892275095, 0.04339627921581268, 0.027448344975709915, 0.09626748412847519, 0.12528061866760254, 0.0906110554933548, 0.016379225999116898, -0.08222322165966034, -0.005582861602306366, -0.02452687919139862, 0.034268636256456375, -0.12091102451086044, 0.01733323372900486, 0.10890744626522064, 0.044852886348962784, 0.10842382907867432, 0.015494686551392078, -0.08181609958410263, 0.003934202715754509, 0.0374157540500164, -0.09569116681814194, -0.17743970453739166, -0.0045800297521054745, 0.021420059725642204, -0.14439913630485535, 0.02729582041501999, 0.0962398499250412, -0.02775685116648674, -0.012751396745443344, -0.009468934498727322, 0.051972895860672, -0.000006967083663766971, 0.15873786807060242, 0.028870906680822372, 0.08189777284860611, -0.060653265565633774, 0.1338346302509308, 0.0966089591383934, -0.12115000933408737, 0.06502479314804077, 0.07614760100841522, -0.08876337856054306, -0.03330419957637787, 0.01473957858979702, 0.06239651143550873, -0.001478298450820148, -0.04186924174427986, -0.03491293638944626, -0.08005660027265549, 0.05378325656056404, 0.039112646132707596, 0.023988477885723114, -0.007630151230841875, 0.010876510292291641, 0.030543260276317596, -0.1327291578054428, 0.11509711295366287, 0.038127340376377106, 0.06738316267728806, -0.1398211568593979, 0.06690353155136108, 0.0015266393311321735, 0.047743696719408035, -0.005486977286636829, 0.017282268032431602, -0.09500119090080261, -0.03462187945842743, -0.13771137595176697, 0.000058130844990955666, -0.060357626527547836, 0.021615738049149513, -0.026182807981967926, -0.04267532378435135, -0.020495600998401642, 0.03508691489696503, -0.05447245016694069, -0.07989078015089035, -0.02053789608180523, 0.07667519897222519, -0.13352586328983307, -0.006133576855063438, 0.032561108469963074, -0.11365827918052673, 0.11846309900283813, 0.030213985592126846, 0.03494831547141075, 0.004002853762358427, -0.05274233594536781, 0.030559532344341278, -0.01135973446071148, 0.042444758117198944, 0.04028394818305969, -0.11117860674858093, -0.0032964921556413174, -0.042731594294309616, -0.005515582859516144, 0.01834745891392231, 0.039090175181627274, -0.1088799387216568, 0.008610272780060768, -0.0606837198138237, -0.04266045242547989, -0.05458515137434006, 0.0494367815554142, 0.07406799495220184, -0.02821209654211998, 0.09513837844133377, -0.058896955102682114, 0.04698210582137108, -0.21514193713665009, -0.04054737836122513, 0.01613817736506462, 0.022557616233825684, -0.031939730048179626, -0.01438306923955679, 0.08481556177139282, -0.037940509617328644, 0.14292895793914795, -0.06701531261205673, 0.07665584236383438, 0.029427150264382362, -0.04989264905452728, -0.0003507330548018217, 0.015832578763365746, 0.16618716716766357, 0.07251261174678802, -0.014221311546862125, 0.051874540746212006, -0.029064228758215904, 0.044674597680568695, 0.003484595799818635, 0.11395333707332611, 0.133119598031044, 0.0004110243171453476, 0.04487074911594391, 0.05963435396552086, -0.1578647941350937, -0.14648225903511047, 0.1425936222076416, -0.07163285464048386, 0.09752768278121948, -0.06576026976108551, 0.10539814829826355, 0.11420192569494247, -0.16924068331718445, 0.03859185054898262, -0.02955225110054016, -0.10844478756189346, -0.09679635614156723, -0.11238692700862885, -0.08045680820941925, -0.10409422218799591, 0.03011292777955532, -0.1183214858174324, 0.030061133205890656, 0.04296408221125603, 0.026099346578121185, 0.016950698569417, 0.14677363634109497, -0.01949325203895569, 0.016019541770219803, 0.05842653661966324, 0.030983423814177513, -0.013314676471054554, -0.007890363223850727, -0.04142403230071068, 0.04865353927016258, 0.010569755919277668, 0.08531016111373901, -0.009043947793543339, 0.0008437783108092844, 0.05260670557618141, 0.008889236487448215, -0.06041208282113075, 0.010429579764604568, -0.0039915479719638824, 0.011669727973639965, 0.08039572089910507, 0.047062475234270096, -0.0012324218405410647, -0.045438334345817566, 0.2535269856452942, -0.05621308460831642, -0.07573935389518738, -0.1362132430076599, 0.12424734234809875, 0.013913247734308243, -0.0020705482456833124, 0.06560664623975754, -0.11401268094778061, 0.0031990609131753445, 0.13833260536193848, 0.15056568384170532, -0.045461367815732956, -0.01920190267264843, 0.02376752905547619, -0.01369452103972435, -0.037890609353780746, 0.10769715160131454, 0.07623981684446335, 0.0812699943780899, -0.053631845861673355, -0.015274162404239178, 0.009445764124393463, -0.035827506333589554, -0.07627064734697342, 0.08658318966627121, 0.004086535889655352, 0.023928629234433174, -0.02649044618010521, 0.07814512401819229, -0.030185135081410408, -0.16738098859786987, 0.019654497504234314, -0.1534615457057953, -0.1893903613090515, -0.02068185992538929, 0.08168122172355652, -0.020724063739180565, 0.04852595552802086, 0.007624490186572075, -0.012664195150136948, 0.15757623314857483, -0.003094869200140238, -0.06258800625801086, -0.06479877233505249, 0.05755484849214554, 0.002048397669568658, 0.23854491114616394, -0.004160646814852953, 0.06490644067525864, 0.1230391412973404, -0.0038409687113016844, -0.16956061124801636, 0.014138760045170784, 0.10087994486093521, -0.06825751811265945, 0.03485826775431633, 0.15926340222358704, -0.03269875794649124, 0.11158829182386398, 0.06997867673635483, -0.06764434278011322, -0.01632579043507576, -0.0035390995908528566, 0.002570557640865445, -0.09512089192867279, 0.03755569830536842, -0.05356592684984207, 0.1817387044429779, 0.16069433093070984, -0.05808253958821297, -0.009090443141758442, -0.052535705268383026, 0.03306744247674942, 0.016652388498187065, 0.12121788412332535, -0.008653904311358929, -0.20258761942386627, 0.0477314218878746, -0.033845674246549606, 0.07293694466352463, -0.21662868559360504, -0.09246900677680969, 0.07718479633331299, -0.02243850938975811, -0.07137852907180786, 0.1323920041322708, 0.09008863568305969, 0.012095184065401554, -0.048833008855581284, -0.16150939464569092, -0.030442094430327415, 0.12115699052810669, -0.15177743136882782, -0.05351820960640907 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
null
imsanjoykb/QLoRA-Finetuning-Barbarik
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
2024-02-10T13:07:46+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 31, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.06646376848220825, 0.2168014943599701, -0.00225935154594481, 0.023818302899599075, 0.1271018385887146, -0.001635765191167593, 0.04218708351254463, 0.13324736058712006, -0.020175931975245476, 0.11144465953111649, 0.046588581055402756, 0.09377603232860565, 0.09928803145885468, 0.18404334783554077, 0.04859916493296623, -0.2059975117444992, 0.007056170143187046, -0.09090408682823181, 0.014076028019189835, 0.1116579994559288, 0.13719257712364197, -0.10291384905576706, 0.08272874355316162, -0.04045208916068077, -0.02019004337489605, 0.00012576708104461432, -0.09259183704853058, -0.07032395154237747, 0.06885425746440887, 0.06264153122901917, 0.051234472543001175, 0.001456156256608665, 0.09140396863222122, -0.2864592671394348, 0.017265573143959045, 0.08406311273574829, 0.0027674848679453135, 0.06290827691555023, 0.07236549258232117, -0.07389893382787704, 0.11328595131635666, -0.08021481335163116, 0.13019037246704102, 0.08625296503305435, -0.062064990401268005, -0.23071379959583282, -0.07525765895843506, 0.0963398814201355, 0.12251301854848862, 0.06215599179267883, -0.022921854630112648, 0.15455181896686554, -0.06248689442873001, 0.012971068732440472, 0.1294165402650833, -0.11526761949062347, -0.05572471022605896, 0.061741601675748825, 0.11775490641593933, 0.10740239918231964, -0.14110268652439117, -0.0017287094378843904, 0.04900608956813812, 0.029121357947587967, 0.08589313924312592, 0.022661056369543076, 0.12003941088914871, 0.04652795568108559, -0.13695219159126282, -0.04037507623434067, 0.12011898308992386, 0.038862764835357666, -0.06446044892072678, -0.2168138176202774, -0.006778308190405369, -0.0601806715130806, -0.014732478186488152, -0.07019448280334473, 0.039128515869379044, -0.02470310963690281, 0.07317749410867691, -0.04465159401297569, -0.1063927412033081, -0.0421026237308979, 0.0892222449183464, 0.07748593389987946, 0.011527054943144321, -0.02519804798066616, 0.04627908393740654, 0.13455867767333984, 0.05402068421244621, -0.10399353504180908, -0.07017925381660461, -0.06942764669656754, -0.09420394152402878, -0.04035796597599983, 0.056760527193546295, 0.031942449510097504, 0.02665667235851288, 0.22703726589679718, 0.016653569415211678, 0.04155244305729866, 0.0224777739495039, 0.01032855175435543, 0.043662428855895996, 0.0955500528216362, -0.05303520709276199, -0.15660029649734497, -0.04072032496333122, 0.09077946096658707, -0.0027527001220732927, -0.036689214408397675, -0.03966725245118141, 0.03849169611930847, 0.06843466311693192, 0.13122352957725525, 0.07552056759595871, -0.017929591238498688, -0.04813180863857269, -0.030096933245658875, 0.23523783683776855, -0.1493375599384308, 0.04426715523004532, -0.02271856553852558, -0.01804111897945404, -0.03908449783921242, 0.03597262129187584, 0.022118929773569107, -0.000004518366949923802, 0.09706240892410278, -0.058981191366910934, -0.05378659814596176, -0.10168042778968811, -0.03272576630115509, 0.04088849574327469, -0.013975566253066063, -0.010589460842311382, -0.09025166928768158, -0.09490354359149933, -0.04766594246029854, 0.05537205561995506, -0.05123869329690933, -0.03770573064684868, 0.009465423412621021, -0.08151785284280777, -0.005444355774670839, -0.005417742300778627, 0.10699385404586792, -0.03222226724028587, 0.04445803165435791, -0.027600755915045738, 0.05225523188710213, 0.09919606149196625, 0.031576547771692276, -0.0773419588804245, 0.0561848059296608, -0.22559374570846558, 0.07503069192171097, -0.11481974273920059, 0.04335082694888115, -0.1704932004213333, -0.042439818382263184, 0.005444696638733149, 0.0139949731528759, 0.013206101022660732, 0.12720820307731628, -0.19255615770816803, -0.01654396951198578, 0.13260798156261444, -0.09212633967399597, -0.118110790848732, 0.07884611934423447, -0.029701577499508858, 0.1624738723039627, 0.04682036489248276, -0.027025915682315826, 0.09224298596382141, -0.16434773802757263, -0.07092688232660294, -0.00949116237461567, -0.01727987825870514, 0.12109188735485077, 0.07512219995260239, -0.05991523340344429, 0.046571120619773865, 0.02832140028476715, -0.038078423589468, -0.04424772411584854, -0.050857074558734894, -0.10884185880422592, -0.01070026308298111, -0.08987759798765182, 0.04065500199794769, -0.01250192429870367, -0.07916021347045898, -0.029885273426771164, -0.18612512946128845, -0.0030564051121473312, 0.10038342326879501, 0.0035033065360039473, -0.005652366206049919, -0.08666291832923889, 0.026358824223279953, -0.03112892620265484, -0.008404186926782131, -0.16764774918556213, -0.04399421438574791, 0.046902090311050415, -0.16094985604286194, 0.020117372274398804, -0.06413903087377548, 0.06334125250577927, 0.03641495108604431, -0.05590536445379257, -0.0248766727745533, -0.01730942726135254, 0.011945613659918308, -0.05083848536014557, -0.18994836509227753, -0.056277405470609665, -0.037882111966609955, 0.149809330701828, -0.25956398248672485, 0.032966937869787216, 0.051140617579221725, 0.14649195969104767, 0.00406361510977149, -0.05115427449345589, 0.01429014839231968, -0.05360214412212372, -0.054652128368616104, -0.06746816635131836, -0.006135428790003061, -0.027576493099331856, -0.05147203803062439, 0.019243421033024788, -0.1755700707435608, -0.021410830318927765, 0.09424154460430145, 0.12876708805561066, -0.1486445665359497, -0.018640631809830666, -0.048725154250860214, -0.06339836865663528, -0.0715010017156601, -0.07038594037294388, 0.10712739825248718, 0.0513901449739933, 0.04796046018600464, -0.07435787469148636, -0.07092321664094925, 0.02726263552904129, 0.006906150374561548, -0.03382374346256256, 0.08727246522903442, 0.05199531093239784, -0.09209315478801727, 0.0756213590502739, 0.1092359870672226, 0.07177663594484329, 0.09363535046577454, 0.01574566215276718, -0.11756632477045059, -0.028492970392107964, 0.036266472190618515, 0.02740776725113392, 0.1465986967086792, -0.05952361226081848, 0.04016614332795143, 0.04494241625070572, -0.04170418903231621, 0.022319864481687546, -0.08787637203931808, 0.024075502529740334, 0.025203049182891846, -0.0034381982404738665, 0.06284574419260025, -0.02525499276816845, -0.0050758360885083675, 0.07016654312610626, 0.047779910266399384, 0.04621000960469246, 0.009655474685132504, -0.01720241829752922, -0.1047825813293457, 0.16950392723083496, -0.0951867327094078, -0.269941508769989, -0.17632324993610382, 0.026197833940386772, 0.04035249724984169, -0.022378476336598396, 0.031619444489479065, -0.07056326419115067, -0.10630585998296738, -0.1060405746102333, -0.002429972169920802, 0.01714223250746727, -0.06364088505506516, -0.0741225928068161, 0.07348573952913284, 0.04382912442088127, -0.14902326464653015, 0.038552410900592804, 0.055694397538900375, -0.057955220341682434, -0.0233661737293005, 0.09118817001581192, 0.12397737801074982, 0.14583967626094818, -0.021366750821471214, -0.028626007959246635, 0.029004426673054695, 0.19620531797409058, -0.13469526171684265, 0.10371150821447372, 0.13814030587673187, -0.04545360431075096, 0.08360563963651657, 0.1560150384902954, 0.029186224564909935, -0.08317049592733383, 0.05044832453131676, 0.04082648828625679, -0.043159641325473785, -0.2666129767894745, -0.0534592866897583, 0.012832709588110447, -0.06255637854337692, 0.09786593168973923, 0.10183793306350708, 0.11542957276105881, 0.034910861402750015, -0.07166364789009094, -0.043925940990448, -0.0058974819257855415, 0.11737963557243347, -0.05490213260054588, -0.012639665976166725, 0.07686592638492584, -0.05086168646812439, 0.005355054512619972, 0.10266812145709991, 0.02973790094256401, 0.17442677915096283, 0.020399179309606552, 0.11231429129838943, 0.06195578724145889, 0.08633565157651901, 0.0007386076031252742, 0.02951662428677082, 0.05147615820169449, 0.017203815281391144, -0.002300140680745244, -0.10421168059110641, -0.006156572140753269, 0.1449710875749588, 0.028103826567530632, 0.029669636860489845, -0.0018948549404740334, -0.005003341939300299, 0.05121048167347908, 0.1746254414319992, -0.011592294089496136, -0.22072425484657288, -0.0845772922039032, 0.06936841458082199, -0.06218599155545235, -0.12968985736370087, -0.026130788028240204, 0.045467354357242584, -0.17519839107990265, 0.026703642681241035, -0.027433741837739944, 0.0919293761253357, -0.09345759451389313, -0.02221956104040146, 0.03687324374914169, 0.084866963326931, -0.014529162086546421, 0.08703910559415817, -0.14498743414878845, 0.11886418610811234, 0.02978132851421833, 0.09024628251791, -0.11081171780824661, 0.07909037172794342, -0.007550720125436783, 0.009180475026369095, 0.19379350543022156, -0.011335089802742004, -0.03514958545565605, -0.08774717897176743, -0.11210042238235474, -0.013537433929741383, 0.12687496840953827, -0.1243172138929367, 0.08773399889469147, -0.015198243781924248, -0.044079482555389404, 0.00937260314822197, -0.12100647389888763, -0.17273177206516266, -0.19628387689590454, 0.05585884302854538, -0.09575839340686798, 0.025643249973654747, -0.11914430558681488, -0.07089093327522278, -0.02952558360993862, 0.241120383143425, -0.1745356321334839, -0.06510113179683685, -0.1468164622783661, -0.046294767409563065, 0.1662203073501587, -0.04437198117375374, 0.0718095526099205, -0.0208172257989645, 0.20345525443553925, 0.005988610442727804, -0.004939318168908358, 0.06724198162555695, -0.08892562240362167, -0.16873881220817566, -0.06771010160446167, 0.1510489284992218, 0.11680185794830322, 0.04907919466495514, -0.002248800592496991, 0.0011772146681323647, -0.016943959519267082, -0.1137804463505745, -0.0033210667315870523, 0.16037839651107788, 0.03878779336810112, 0.025986969470977783, -0.05243593826889992, -0.08797456324100494, -0.06899320334196091, -0.06853509694337845, 0.06221301481127739, 0.19590823352336884, -0.10376439243555069, 0.1700313836336136, 0.147536963224411, -0.07305635511875153, -0.23175598680973053, 0.035342130810022354, 0.04983805492520332, 0.0014306638622656465, 0.04886869341135025, -0.18252557516098022, 0.10521943867206573, 0.019543392583727837, -0.05505957826972008, 0.13485197722911835, -0.1557481735944748, -0.1552847921848297, 0.0722852572798729, 0.03904085233807564, -0.22423844039440155, -0.1354004591703415, -0.09622503817081451, -0.05825018882751465, -0.14065024256706238, 0.06054598465561867, -0.002136280992999673, 0.015948504209518433, 0.03500790148973465, -0.0015643214574083686, 0.027123261243104935, -0.058935679495334625, 0.18609118461608887, -0.004065449349582195, 0.020676052197813988, -0.060264769941568375, -0.0478842556476593, 0.09839435666799545, -0.06130504235625267, 0.12208222597837448, 0.004057085141539574, 0.01594383642077446, -0.10362856835126877, -0.048314861953258514, -0.04328322783112526, 0.05154227837920189, -0.07548051327466965, -0.10070807486772537, -0.043625857681035995, 0.08841723203659058, 0.07005169242620468, -0.03383097052574158, 0.00549331633374095, -0.07189501076936722, 0.10019614547491074, 0.17795267701148987, 0.17573626339435577, 0.009926567785441875, -0.07241068035364151, 0.01677953451871872, -0.04142116755247116, 0.044231921434402466, -0.2513144314289093, 0.03756171092391014, 0.06098250672221184, 0.029438555240631104, 0.09217222779989243, -0.020435843616724014, -0.1820858269929886, -0.04050002992153168, 0.08094815909862518, -0.05452597141265869, -0.22617179155349731, -0.019085140898823738, 0.0954197570681572, -0.2020406424999237, -0.007372708059847355, 0.03995226323604584, -0.048725228756666183, -0.023169852793216705, 0.00010950004070764408, 0.06317184865474701, 0.002471912419423461, 0.09773622453212738, 0.0735151618719101, 0.09715340286493301, -0.08337292820215225, 0.10562895983457565, 0.10150538384914398, -0.09572599828243256, 0.03605884686112404, 0.06754924356937408, -0.05300498008728027, -0.043293699622154236, 0.03665391728281975, 0.033023297786712646, 0.005234600510448217, -0.060321882367134094, 0.013913018628954887, -0.036497246474027634, 0.044923391193151474, 0.08326134830713272, 0.03754979372024536, -0.013354414142668247, 0.06462216377258301, 0.03401726484298706, -0.10898099094629288, 0.10366570204496384, 0.01731540448963642, 0.04105307161808014, -0.08384523540735245, -0.019968897104263306, 0.035425446927547455, 0.030576206743717194, -0.01765924133360386, -0.02306121215224266, -0.02860277332365513, -0.01614218018949032, -0.14299540221691132, -0.023106401786208153, -0.07243485748767853, 0.006181265693157911, 0.014656842686235905, -0.031884219497442245, -0.011233693920075893, 0.02475680410861969, -0.06979699432849884, -0.07426341623067856, -0.006949664559215307, 0.09833318740129471, -0.15115703642368317, 0.008848577737808228, 0.06907843053340912, -0.11088496446609497, 0.08190931379795074, -0.008411259390413761, 0.016245156526565552, 0.022527478635311127, -0.15448406338691711, 0.05601610988378525, 0.0008648968650959432, 0.01916889287531376, 0.025886621326208115, -0.16471809148788452, 0.004104440100491047, -0.04661374166607857, -0.02149827405810356, -0.00004464812809601426, -0.02647159807384014, -0.12325995415449142, 0.06858719140291214, -0.015622655861079693, -0.035931166261434555, -0.02701525390148163, 0.0539589487016201, 0.07888586074113846, -0.027474910020828247, 0.10445091128349304, -0.008690856397151947, 0.04941811040043831, -0.16801609098911285, -0.02470702864229679, -0.04982255399227142, 0.019377702847123146, 0.009884213097393513, -0.007693959400057793, 0.04183054715394974, -0.00976533442735672, 0.21883612871170044, -0.05075952783226967, 0.1607085019350052, 0.05847611650824547, -0.017352959141135216, -0.0007513365126214921, 0.06180921941995621, 0.05997028574347496, 0.04658793285489082, 0.009480604901909828, 0.023740366101264954, -0.022450892254710197, -0.006695089396089315, -0.15932634472846985, 0.01890849508345127, 0.14999441802501678, 0.06301083415746689, 0.024745315313339233, 0.05866100639104843, -0.12775006890296936, -0.12135478109121323, 0.09311001747846603, -0.026755332946777344, 0.00928465835750103, -0.08245618641376495, 0.1358020007610321, 0.14980104565620422, -0.14000412821769714, 0.05256148427724838, -0.06134212389588356, -0.05217423290014267, -0.10388828068971634, -0.12032219022512436, -0.05887215584516525, -0.053666237741708755, 0.002330566756427288, -0.03760887682437897, 0.054546963423490524, 0.03344334661960602, -0.009351172484457493, -0.00022941511997487396, 0.13597318530082703, -0.019751882180571556, -0.0028988157864660025, 0.048313532024621964, 0.03693558648228645, 0.02373051457107067, -0.05275435373187065, 0.02940409444272518, 0.02539868652820587, 0.032232340425252914, 0.06546790152788162, 0.033412106335163116, -0.047448933124542236, 0.03804153576493263, -0.0025254099164158106, -0.11207924783229828, 0.019641218706965446, -0.00460948096588254, -0.0742158442735672, 0.1268945336341858, 0.0407399944961071, 0.010224059224128723, -0.03741471841931343, 0.24361543357372284, -0.06653323769569397, -0.06378097087144852, -0.13251738250255585, 0.10491154342889786, -0.0027236645109951496, 0.06476365029811859, 0.023412218317389488, -0.1284150779247284, 0.005243356805294752, 0.13858191668987274, 0.12181595712900162, 0.0045748427510261536, 0.009228081442415714, 0.0518609918653965, 0.0025186820421367884, -0.06998204439878464, 0.054019294679164886, 0.06992026418447495, 0.12919506430625916, -0.07847554981708527, 0.07680778950452805, 0.0006860480643808842, -0.08370215445756912, -0.02947772853076458, 0.11312682181596756, -0.0409729965031147, 0.03491825982928276, -0.047444481402635574, 0.10916327685117722, -0.05787910893559456, -0.29412412643432617, 0.02350960113108158, -0.09588567912578583, -0.15202060341835022, -0.018367812037467957, 0.05944539234042168, -0.02624768204987049, 0.018029648810625076, 0.06971040368080139, -0.06011629104614258, 0.20098382234573364, 0.0335683599114418, -0.07864278554916382, -0.0664360448718071, 0.04837050288915634, -0.06564252078533173, 0.2949807047843933, 0.008418165147304535, 0.02863333560526371, 0.10770907253026962, -0.03253700211644173, -0.18271861970424652, 0.010723991319537163, 0.1133992001414299, -0.08056149631738663, 0.08200647681951523, 0.19000613689422607, -0.012578671798110008, 0.1209007054567337, 0.05294662341475487, -0.047376248985528946, 0.04217283055186272, -0.03389401361346245, -0.051268599927425385, -0.10752558708190918, 0.058453381061553955, -0.05909625440835953, 0.15447644889354706, 0.10152646154165268, -0.05671518296003342, -0.004550917539745569, -0.05555408447980881, 0.04875178262591362, 0.01804669201374054, 0.12263146042823792, 0.02951994352042675, -0.1865430772304535, 0.032826557755470276, -0.01144319772720337, 0.10186848044395447, -0.25588861107826233, -0.08421015739440918, 0.08833149075508118, -0.011924264021217823, -0.05105875805020332, 0.10560628771781921, 0.057650718837976456, 0.04243382066488266, -0.043439045548439026, -0.10480839014053345, -0.02186836116015911, 0.14663739502429962, -0.1469624787569046, -0.025013303384184837 ]
null
null
null
GGUF quant of [alexshengzhili/llava-v1.5-13b-dpo](https://huggingface.co/alexshengzhili/llava-v1.5-13b-dpo)
{"license": "llama2", "tags": ["gguf", "dpo"]}
null
antiven0m/llava-v1.5-13b-dpo-gguf
[ "gguf", "dpo", "license:llama2", "region:us" ]
2024-02-10T13:10:18+00:00
[]
[]
TAGS #gguf #dpo #license-llama2 #region-us
GGUF quant of alexshengzhili/llava-v1.5-13b-dpo
[]
[ "TAGS\n#gguf #dpo #license-llama2 #region-us \n" ]
[ 19 ]
[ "passage: TAGS\n#gguf #dpo #license-llama2 #region-us \n" ]
[ -0.008604014292359352, 0.10990592837333679, -0.010436690412461758, -0.011385667137801647, 0.029359333217144012, 0.038117341697216034, 0.16163727641105652, 0.04485592618584633, 0.17062106728553772, -0.03827083483338356, 0.1400679349899292, 0.10981146991252899, 0.03900774568319321, -0.013631194829940796, -0.014640500769019127, -0.19766303896903992, 0.054711345583200455, -0.0447491779923439, -0.027350930497050285, 0.013151803985238075, 0.023680707439780235, -0.03657340630888939, 0.01089315302670002, -0.04156934842467308, -0.08463115245103836, 0.01594356633722782, 0.006978940684348345, -0.041349221020936966, 0.08800700306892395, 0.06240101531147957, 0.07438506931066513, 0.0690712183713913, -0.014287578873336315, -0.19010889530181885, 0.019369730725884438, -0.07039441913366318, -0.1017642617225647, 0.00999301578849554, 0.03415362909436226, 0.03626423701643944, 0.08497247844934464, 0.10314887762069702, -0.05619346722960472, 0.007611309178173542, -0.1933366060256958, -0.22296294569969177, -0.0822427049279213, 0.03789922967553139, 0.0150843420997262, 0.037769995629787445, 0.0753028616309166, 0.07307632267475128, -0.16651317477226257, -0.032483261078596115, 0.06368890404701233, -0.25846725702285767, 0.060304947197437286, 0.22062835097312927, 0.026245877146720886, 0.006638676393777132, -0.012579119764268398, 0.10754617303609848, 0.01451259944587946, -0.05401061847805977, -0.1630697101354599, -0.055702392011880875, 0.07093190401792526, 0.12427330762147903, -0.029690369963645935, -0.05375698208808899, 0.22310078144073486, 0.04175828397274017, -0.0285833477973938, 0.09888460487127304, 0.005592044442892075, 0.004661146551370621, 0.01614573411643505, -0.03722082078456879, 0.06555384397506714, 0.16337214410305023, 0.16595213115215302, -0.01352343987673521, -0.1162264421582222, -0.040554363280534744, -0.2636548578739166, 0.19429932534694672, -0.007079020142555237, 0.09835553169250488, -0.18632540106773376, 0.023846335709095, -0.17257249355316162, -0.011174255982041359, -0.08603711426258087, -0.022913886234164238, 0.12061678618192673, 0.004377703182399273, -0.018880898132920265, 0.1097363829612732, 0.16012230515480042, 0.09268757700920105, -0.009057335555553436, -0.024257777258753777, -0.03067002072930336, 0.15153947472572327, 0.0009907269850373268, 0.037922490388154984, 0.05955795571208, 0.16124281287193298, 0.03338342532515526, -0.17384476959705353, 0.04011434689164162, -0.007287521846592426, -0.1236291378736496, 0.008939811959862709, -0.14288049936294556, 0.1409846395254135, -0.01718830317258835, -0.054926831275224686, -0.04788605123758316, 0.02268046885728836, 0.08396497368812561, -0.026906471699476242, -0.04215921461582184, -0.02619314007461071, -0.010400473140180111, -0.024827243760228157, -0.07189614325761795, 0.05608661100268364, 0.11880996078252792, 0.11779796332120895, -0.11855059862136841, -0.0310702845454216, 0.019885238260030746, 0.07338982820510864, 0.09219211339950562, -0.04314781725406647, 0.08962536603212357, -0.12877598404884338, -0.15621456503868103, 0.020945502445101738, -0.010420521721243858, -0.01878209412097931, 0.03551579639315605, 0.05456192418932915, 0.04903218150138855, -0.06990807503461838, -0.06104813516139984, -0.07332301139831543, -0.05493337661027908, 0.1324758529663086, -0.0028390884399414062, -0.006768906954675913, -0.26182055473327637, 0.009011900052428246, -0.05480508133769035, 0.06569038331508636, -0.002690385328605771, -0.052001453936100006, -0.07638682425022125, 0.13771317899227142, 0.01270152349025011, 0.060601141303777695, -0.12614652514457703, 0.017068572342395782, -0.056371502578258514, 0.12068279832601547, -0.08770890533924103, -0.0817338302731514, 0.14877544343471527, -0.11088868975639343, -0.07305719703435898, -0.001013837056234479, 0.03150518983602524, 0.036342181265354156, 0.0878019779920578, 0.34138038754463196, -0.02566511183977127, -0.22543172538280487, 0.07243802398443222, 0.14366744458675385, -0.1592685431241989, -0.19014078378677368, 0.14730755984783173, -0.15642289817333221, -0.07279204577207565, 0.038588788360357285, -0.04594062268733978, 0.11595360934734344, -0.041179314255714417, -0.070935919880867, 0.04900781065225601, -0.02434147149324417, -0.07041450589895248, 0.030703015625476837, 0.0667063370347023, -0.043675851076841354, 0.06523914635181427, -0.013068880885839462, 0.04957098513841629, 0.14224180579185486, -0.04011807590723038, -0.08189385384321213, 0.03933168575167656, 0.033194709569215775, 0.011906391009688377, 0.025627465918660164, -0.13208061456680298, -0.01836722530424595, -0.014745086431503296, 0.044013094156980515, 0.1125682145357132, 0.05061475187540054, -0.03446149453520775, 0.013545257970690727, 0.0862707644701004, 0.041738949716091156, 0.042152903974056244, 0.041745640337467194, -0.03473641350865364, 0.12255382537841797, -0.016534289345145226, -0.00947380531579256, -0.05617741495370865, -0.02797873318195343, 0.18606658279895782, -0.03801412880420685, 0.0016870887484401464, -0.0025432512629777193, -0.018611550331115723, -0.024748438969254494, 0.08651616424322128, -0.008789781481027603, 0.12290797382593155, 0.026027293875813484, -0.05689841881394386, 0.14643293619155884, 0.019987374544143677, 0.3014521598815918, 0.15490050613880157, -0.01079521793872118, -0.008067501708865166, -0.11539771407842636, -0.030958127230405807, 0.01649365946650505, 0.03937772661447525, 0.0032656665425747633, -0.018702050670981407, -0.06289736926555634, -0.006221037823706865, -0.006192989181727171, 0.010551425628364086, -0.013262656517326832, -0.01514099445194006, -0.10309090465307236, 0.041045986115932465, 0.21785926818847656, -0.13373206555843353, 0.15760298073291779, 0.280707985162735, 0.15049955248832703, 0.19441139698028564, -0.1364663988351822, -0.01611248403787613, -0.06313829869031906, 0.054778244346380234, -0.027937451377511024, 0.17334116995334625, -0.07432560622692108, 0.03158557415008545, 0.03421064838767052, 0.003000217257067561, 0.04942243546247482, -0.15744920074939728, -0.19559864699840546, -0.02506384812295437, -0.05773289129137993, -0.1618327498435974, 0.09337372332811356, -0.11417675018310547, 0.04794899374246597, 0.05846836417913437, -0.12129863351583481, 0.16077010333538055, -0.0025180026423186064, -0.02169257029891014, 0.0546196848154068, -0.11778664588928223, -0.13620276749134064, -0.02523024193942547, -0.02548091858625412, -0.009553107433021069, 0.05594167858362198, 0.03854655846953392, -0.08925442397594452, -0.0013303147861734033, 0.12127372622489929, -0.009187128394842148, -0.1576712280511856, 0.01721201464533806, 0.021827921271324158, 0.06792574375867844, -0.07916224747896194, -0.08761003613471985, -0.08766653388738632, -0.08618664741516113, -0.020556364208459854, 0.07689250260591507, -0.03411305323243141, 0.060171473771333694, 0.12044026702642441, 0.08443263173103333, 0.08545760810375214, -0.0218313280493021, 0.11368317902088165, -0.07866497337818146, -0.12182459980249405, 0.09807919710874557, 0.03870660439133644, 0.027706263586878777, 0.16150686144828796, 0.09919442981481552, -0.15118730068206787, -0.01451702881604433, -0.026812037453055382, -0.1099858507514, -0.15622341632843018, -0.03067006729543209, -0.06864404678344727, 0.1355171650648117, -0.0726400762796402, 0.10953298956155777, 0.0953807607293129, 0.04051729291677475, 0.09226501733064651, -0.05300109460949898, -0.01567668654024601, -0.01864701695740223, 0.2293231040239334, -0.05454541742801666, -0.025683553889393806, -0.10517170280218124, -0.0725967288017273, 0.13389645516872406, 0.12586544454097748, 0.08336679637432098, 0.2467498779296875, 0.0034341581631451845, 0.1152256652712822, 0.04626575857400894, 0.09000218659639359, -0.028399812057614326, 0.06476333737373352, -0.04814096540212631, -0.037126414477825165, -0.027179427444934845, 0.019886499270796776, 0.02630574256181717, 0.04715466871857643, -0.21906906366348267, 0.03085169568657875, -0.2538618743419647, -0.014681653119623661, -0.053767506033182144, 0.042159028351306915, 0.02630297839641571, 0.07437022775411606, 0.0682765319943428, 0.05077698081731796, 0.017237048596143723, 0.10215983539819717, -0.0664747878909111, -0.08377718925476074, 0.027363600209355354, 0.07601329684257507, 0.01458102185279131, -0.003094865009188652, 0.05726925656199455, -0.07787586003541946, -0.12715330719947815, 0.0393931046128273, 0.1057836264371872, -0.2990707755088806, 0.2321292906999588, 0.033978331834077835, -0.06051423400640488, -0.044214725494384766, -0.04396810755133629, 0.006165951490402222, 0.09439336508512497, 0.15290741622447968, 0.06583677232265472, 0.008076618425548077, -0.06522484868764877, -0.01193122286349535, 0.030653364956378937, 0.0868416279554367, -0.05719485506415367, -0.16691574454307556, -0.006546330638229847, 0.05637882649898529, 0.01664353907108307, 0.09064028412103653, -0.09170889109373093, -0.029484810307621956, 0.08025582134723663, 0.038538750261068344, 0.08386664092540741, -0.05773128196597099, 0.02831965684890747, -0.11404731869697571, 0.07285003364086151, -0.02676057256758213, -0.017409875988960266, -0.06883085519075394, -0.1343785673379898, -0.050361406058073044, -0.049571700394153595, -0.015976129099726677, -0.02963128872215748, -0.17469871044158936, -0.07923954725265503, -0.17735837399959564, 0.14337241649627686, -0.02449001930654049, 0.049254585057497025, -0.0617779865860939, 0.15044823288917542, -0.029538193717598915, 0.00392355490475893, -0.0205406341701746, 0.000011974055269092787, 0.028038358315825462, -0.13013166189193726, 0.13609398901462555, -0.07737501710653305, -0.007063171360641718, -0.003404991701245308, 0.022458195686340332, 0.13208214938640594, 0.05983384698629379, -0.1283237785100937, 0.16550718247890472, 0.3179059624671936, -0.00009329552267445251, 0.20990535616874695, 0.17456696927547455, -0.07864280045032501, -0.21234120428562164, -0.145003542304039, -0.1986551731824875, -0.0725696012377739, 0.053191863000392914, -0.1863056719303131, 0.02324720099568367, 0.23041310906410217, -0.11910489201545715, 0.3675578236579895, -0.29636481404304504, -0.037297796458005905, 0.17703364789485931, -0.035510141402482986, 0.5763704180717468, -0.18657514452934265, -0.1216515526175499, 0.016097385436296463, -0.21444196999073029, 0.21344783902168274, 0.021977657452225685, 0.05891115590929985, -0.01677001267671585, -0.0014102024724707007, 0.030207550153136253, -0.002529863500967622, 0.21785740554332733, 0.02860812470316887, 0.05936092883348465, -0.10149716585874557, -0.11197908967733383, 0.11004850268363953, 0.05216795578598976, -0.020141012966632843, 0.012627960182726383, -0.04067001864314079, -0.10046157985925674, -0.021544629707932472, -0.03920109197497368, 0.07546880841255188, 0.03680989146232605, -0.11902809143066406, -0.0787229984998703, 0.034105636179447174, -0.12174433469772339, -0.038444653153419495, 0.1800990253686905, -0.048154935240745544, 0.03333315998315811, -0.016548877581954002, -0.07953057438135147, -0.10433286428451538, -0.025705907493829727, -0.09812047332525253, -0.04440317302942276, 0.09789538383483887, -0.10573495924472809, -0.027584858238697052, 0.10564152896404266, 0.023615391924977303, 0.041913796216249466, 0.07522698491811752, -0.08149159699678421, 0.01515471376478672, 0.13297149538993835, -0.12931565940380096, -0.15441924333572388, 0.0028588653076440096, -0.06345749646425247, 0.23164716362953186, 0.05686921253800392, 0.025887077674269676, 0.05352102592587471, 0.02621527947485447, 0.03213569521903992, 0.02946140430867672, -0.12739278376102448, -0.019575197249650955, 0.07319095730781555, -0.042914073914289474, -0.10538321733474731, 0.13602103292942047, 0.0541912280023098, 0.01610880345106125, -0.05930979922413826, 0.06908126175403595, -0.019835980609059334, -0.041671328246593475, -0.27050265669822693, 0.015680033713579178, -0.17881007492542267, -0.07175756245851517, 0.020230967551469803, -0.10003606230020523, 0.003320024348795414, 0.049053896218538284, 0.021847855299711227, 0.16767176985740662, 0.04368465393781662, 0.009624972008168697, 0.08005034178495407, -0.07546933740377426, -0.18326197564601898, -0.0024468847550451756, -0.11745510995388031, -0.09636630117893219, -0.03108089044690132, 0.1031213328242302, -0.06991948932409286, -0.08096121996641159, -0.21689341962337494, 0.05969030037522316, -0.06589128822088242, -0.09015432745218277, -0.053781747817993164, -0.011442683637142181, 0.0511784590780735, -0.05011094734072685, 0.007357196416705847, -0.015386167913675308, -0.13423462212085724, 0.03824406862258911, 0.06612880527973175, 0.05289808660745621, -0.03629519045352936, -0.07660531252622604, 0.07769761234521866, 0.05614930018782616, 0.12855671346187592, 0.1134600043296814, 0.0912487581372261, 0.10416565835475922, -0.15870563685894012, 0.00989968329668045, 0.12234783917665482, -0.037433404475450516, -0.0146408686414361, 0.050973035395145416, -0.029200492426753044, 0.04349001497030258, -0.07275420427322388, 0.07256379723548889, -0.04815380647778511, -0.14213241636753082, -0.09233657270669937, 0.02407810464501381, -0.12162399291992188, 0.02072307839989662, -0.1660524159669876, 0.13902193307876587, 0.03309563174843788, 0.03716035932302475, 0.06200183928012848, 0.0002745137317106128, 0.01291372999548912, -0.006037838291376829, 0.0015096059069037437, -0.08083322644233704, 0.012766605243086815, -0.012432161718606949, -0.06893059611320496, -0.007279359735548496, 0.3189520537853241, 0.0031216442584991455, -0.19956175982952118, 0.030055047944188118, 0.08866199105978012, 0.08846539258956909, -0.024563975632190704, 0.24795633554458618, 0.10392896085977554, 0.0037063895724713802, -0.14644363522529602, 0.08284571766853333, -0.01996508426964283, -0.19614766538143158, 0.08459345996379852, 0.03933154419064522, -0.026528291404247284, -0.010751902125775814, 0.11682318896055222, -0.1047138050198555, 0.022269342094659805, -0.05693594738841057, 0.04497012868523598, -0.06619520485401154, -0.011736071668565273, 0.00868336670100689, 0.17678439617156982, -0.009095869958400726, 0.013609562069177628, -0.03582657501101494, 0.024426156654953957, -0.13359050452709198, -0.10515575855970383, 0.04086108133196831, -0.10164874792098999, 0.09123267978429794, 0.02414095588028431, 0.05485517531633377, 0.18281376361846924, 0.01958405040204525, -0.06559038907289505, -0.040410835295915604, -0.10262871533632278, -0.0800689235329628, -0.00900281686335802, -0.02355310693383217, -0.010505080223083496, -0.16096274554729462, -0.07442469894886017, -0.04779975488781929, -0.08344926685094833, -0.06377021223306656, -0.010941260494291782, 0.004069055896252394, -0.06556198745965958, -0.12845826148986816, 0.015858178958296776, -0.05927879363298416, 0.1083315834403038, -0.0118231400847435, 0.16096776723861694, -0.017704574391245842, 0.019217658787965775, 0.06601261347532272, 0.06481006741523743, 0.04007667675614357, -0.05785553529858589, 0.05443761497735977, 0.04309026151895523, -0.012815847061574459, 0.13736487925052643, -0.05632287636399269, -0.0036651426926255226, -0.0020264333579689264, 0.15221676230430603, 0.23317791521549225, -0.06547090411186218, 0.03633522987365723, -0.005018163006752729, 0.022938977926969528, 0.10394994914531708, 0.09812945127487183, 0.04346131905913353, 0.2976507842540741, -0.07777208834886551, -0.01047578826546669, 0.014305406250059605, 0.06951022893190384, -0.08609791845083237, 0.07553587853908539, 0.06683829426765442, -0.049185555428266525, -0.0659327283501625, 0.10126569867134094, -0.15109826624393463, 0.10064199566841125, 0.10488591343164444, -0.11295390874147415, 0.012834766879677773, -0.04432997107505798, -0.05450597032904625, -0.006060434505343437, 0.07373521476984024, -0.109685517847538, -0.0887870341539383, -0.06705191731452942, 0.06362226605415344, -0.37490424513816833, -0.17265357077121735, 0.0707719698548317, 0.22528743743896484, 0.11500155180692673, -0.04423389583826065, 0.10855096578598022, 0.0013076248578727245, 0.07358778268098831, -0.042039014399051666, 0.14836634695529938, 0.03534693643450737, -0.06350690871477127, -0.14247187972068787, -0.20906347036361694, 0.010107136331498623, -0.09563980996608734, 0.048075634986162186, 0.07950121909379959, -0.01766333356499672, 0.1395152509212494, -0.06183941289782524, -0.033935315907001495, 0.05668232589960098, -0.15559810400009155, 0.012556914240121841, -0.030367685481905937, 0.023553799837827682, -0.0386824905872345, -0.0755348950624466, 0.003964729141443968, 0.1255546510219574, -0.21429072320461273, -0.07645464688539505, 0.16537998616695404, 0.004544720519334078, 0.13568337261676788, -0.0321102999150753, -0.022601788863539696, -0.010483797639608383, -0.11098100990056992, 0.09499136358499527, -0.06953500211238861, -0.017798656597733498, 0.1294688880443573, -0.0004968442372046411, 0.009858810342848301, -0.23295891284942627, 0.030766261741518974, -0.0462636724114418, -0.04952159523963928, -0.06220163404941559 ]
null
null
null
# **Q-Learning** Agent playing1 **FrozenLake-v1** This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** . ## Usage ```python model = load_from_hub(repo_id="aturja65/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
{"tags": ["FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "q-FrozenLake-v1-4x4-noSlippery", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "FrozenLake-v1-4x4-no_slippery", "type": "FrozenLake-v1-4x4-no_slippery"}, "metrics": [{"type": "mean_reward", "value": "1.00 +/- 0.00", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
aljaziz/q-FrozenLake-v1-4x4-noSlippery
[ "FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
2024-02-10T13:11:07+00:00
[]
[]
TAGS #FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
# Q-Learning Agent playing1 FrozenLake-v1 This is a trained model of a Q-Learning agent playing FrozenLake-v1 . ## Usage
[ "# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage" ]
[ "TAGS\n#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n", "# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage" ]
[ 40, 39 ]
[ "passage: TAGS\n#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage" ]
[ 0.04578453302383423, -0.08074592798948288, -0.00430759321898222, 0.10720831900835037, 0.05034215748310089, -0.040469273924827576, 0.11997015029191971, 0.018999949097633362, 0.20601962506771088, -0.010012076236307621, 0.1455274522304535, 0.007022971753031015, -0.006192410364747047, 0.1867983490228653, 0.04572829231619835, -0.26324528455734253, 0.01831899583339691, -0.09495259821414948, -0.07281816750764847, 0.11870454251766205, 0.05470194295048714, -0.01901467889547348, -0.0007633853238075972, 0.056141503155231476, -0.0673527717590332, 0.0007737681735306978, 0.031996939331293106, -0.012976245954632759, 0.19804789125919342, -0.02254498563706875, 0.06641989201307297, 0.054705578833818436, 0.0758768692612648, -0.1998077929019928, 0.0358855277299881, -0.04215473681688309, -0.09439758956432343, -0.03934839740395546, -0.018780618906021118, 0.05878105387091637, 0.053356342017650604, 0.03858819976449013, 0.058354366570711136, 0.09384993463754654, -0.0773480236530304, 0.04328357055783272, 0.04280758649110794, 0.024811049923300743, 0.04589218273758888, -0.0237203948199749, -0.027002155780792236, 0.08246652781963348, -0.22182892262935638, 0.10318073630332947, -0.010159241035580635, -0.5270710587501526, -0.00633762264624238, 0.24088262021541595, 0.11517096310853958, 0.05707438662648201, -0.06903956830501556, 0.10566288232803345, 0.03913382440805435, -0.007209456991404295, 0.03210983797907829, 0.02150118350982666, 0.12817370891571045, 0.06009242683649063, -0.09581366181373596, 0.040699947625398636, 0.13722525537014008, 0.012822695076465607, 0.020306183025240898, -0.08888901025056839, 0.0410032719373703, -0.03461858257651329, -0.007679527159780264, -0.09758518636226654, 0.05478060990571976, 0.012466507963836193, -0.0934976264834404, -0.09247440844774246, -0.04236573353409767, -0.06708304584026337, 0.11252415925264359, 0.046419668942689896, -0.0874939113855362, 0.03884070739150047, -0.06760413944721222, 0.05918780341744423, -0.16863860189914703, 0.02074250765144825, -0.06627868115901947, -0.09376336634159088, -0.11799788475036621, -0.01683047041296959, -0.07946427166461945, 0.009092256426811218, 0.056664444506168365, 0.1447116881608963, 0.22076484560966492, 0.06690320372581482, 0.09728849679231644, 0.07456006109714508, 0.06531001627445221, 0.1538129299879074, 0.10918238013982773, 0.019075315445661545, -0.015266558155417442, 0.0948706716299057, -0.06445580720901489, -0.1351388692855835, -0.15579092502593994, 0.005488025024533272, 0.0983937531709671, 0.08871900290250778, -0.044080477207899094, -0.006702381651848555, -0.024641724303364754, 0.08566431701183319, -0.11314457654953003, -0.024612564593553543, -0.002267979085445404, 0.06882024556398392, -0.024801667779684067, 0.020378148183226585, -0.06242705136537552, 0.12715265154838562, 0.04222423583269119, -0.059924717992544174, -0.055308472365140915, -0.03053177334368229, -0.014276440255343914, -0.027539284899830818, 0.02446848154067993, -0.07659092545509338, 0.04767750948667526, -0.16766095161437988, -0.042871296405792236, -0.04784649610519409, 0.025697942823171616, -0.03907240927219391, -0.13557587563991547, -0.17699143290519714, -0.048906855285167694, -0.022438718006014824, 0.03549358621239662, -0.038111843168735504, 0.006551501806825399, -0.006318534724414349, -0.1583600640296936, 0.09783563017845154, 0.09784027189016342, -0.03643378987908363, -0.02749447710812092, 0.056263517588377, -0.07194498926401138, 0.1561182290315628, -0.21054518222808838, -0.054014235734939575, -0.044764336198568344, -0.06595750898122787, 0.19673264026641846, 0.012690845876932144, -0.01202624011784792, 0.19873127341270447, -0.29073721170425415, -0.06078760325908661, 0.12533614039421082, -0.07834373414516449, -0.0936407670378685, 0.06941844522953033, -0.04206686094403267, 0.023345354944467545, 0.046047765761613846, 0.36345911026000977, -0.02069227211177349, -0.16197136044502258, -0.021782705560326576, 0.13971707224845886, -0.1184760183095932, 0.059895481914281845, 0.04240793362259865, 0.12543781101703644, -0.04250509291887283, -0.018672896549105644, -0.09023164212703705, 0.05999075248837471, -0.05241934582591057, -0.09016361832618713, -0.03393383324146271, -0.07645075023174286, 0.13294468820095062, -0.0629684180021286, 0.05601520463824272, -0.03255095332860947, -0.07133250683546066, -0.050324998795986176, -0.016492370516061783, 0.04460815340280533, 0.05951254442334175, -0.12794871628284454, 0.11029167473316193, 0.13025271892547607, -0.0006193425506353378, -0.07498852163553238, -0.17872096598148346, 0.003240168560296297, 0.009576505981385708, 0.039837226271629333, 0.17141658067703247, 0.12209978699684143, 0.033295199275016785, 0.008770671673119068, -0.06389404833316803, -0.18276847898960114, 0.058129217475652695, -0.056212130934000015, -0.14230976998806, -0.052409034222364426, -0.0728459507226944, 0.017381802201271057, -0.0859743058681488, -0.017379917204380035, 0.021926190704107285, 0.006908397190272808, 0.02990424446761608, -0.026645656675100327, -0.049561817198991776, 0.021254703402519226, 0.06490101665258408, -0.0037617047782987356, 0.12023693323135376, 0.008277264423668385, -0.18308481574058533, 0.07930773496627808, 0.08478537946939468, 0.09196605533361435, 0.013250201940536499, 0.02685922384262085, -0.021522263064980507, -0.08061408251523972, -0.054420311003923416, 0.02957955375313759, 0.11417073011398315, 0.1317172348499298, 0.2361993044614792, 0.08753683418035507, 0.04697408527135849, -0.02164587564766407, -0.016415923833847046, 0.002810494042932987, -0.06318057328462601, -0.029935607686638832, 0.10614971816539764, 0.05865858122706413, -0.067733034491539, -0.04576427489519119, 0.09590928256511688, 0.02732124738395214, 0.21205885708332062, -0.03342745825648308, 0.01286078616976738, -0.10957037657499313, -0.06550975888967514, -0.031982194632291794, 0.09201868623495102, 0.09498392790555954, 0.009755023755133152, -0.022056059911847115, -0.04259001836180687, 0.0012916827108711004, -0.1334889680147171, -0.10375088453292847, 0.026475343853235245, 0.013400445692241192, -0.11206940561532974, 0.11674030870199203, -0.11352457851171494, 0.039504457265138626, 0.06024791672825813, -0.13837239146232605, 0.04428480193018913, -0.029713207855820656, -0.07886212319135666, 0.16866780817508698, -0.11075661331415176, -0.094340018928051, -0.08831550180912018, 0.004082420375198126, 0.0075836325995624065, -0.03922267258167267, -0.009283260442316532, -0.19952571392059326, -0.005375816952437162, -0.03544965013861656, 0.013616434298455715, -0.06988783925771713, -0.11287739872932434, -0.010957922786474228, 0.07084179669618607, -0.043388739228248596, -0.07803605496883392, 0.007967432029545307, -0.08923084288835526, -0.10623309016227722, 0.028189711272716522, 0.019765101373195648, -0.022883659228682518, 0.16152891516685486, 0.01816628873348236, 0.05626589432358742, -0.03298520669341087, 0.30665266513824463, -0.038163769990205765, 0.08371731638908386, -0.02993497997522354, -0.07433546334505081, 0.06130730360746384, -0.022327827289700508, 0.06086638569831848, -0.020221687853336334, -0.02362890914082527, 0.0077952733263373375, -0.08579335361719131, -0.18365982174873352, -0.05417544022202492, 0.03724347800016403, 0.195254847407341, 0.031118987128138542, 0.01910330168902874, -0.0488768145442009, -0.010547760874032974, 0.1665220558643341, -0.10005921125411987, 0.04030545800924301, -0.05366240441799164, 0.11506262421607971, -0.08640182018280029, 0.06195629760622978, 0.020486772060394287, 0.04266135022044182, -0.04877188801765442, 0.09486009180545807, 0.0826394334435463, 0.1121082529425621, -0.02206910029053688, 0.046257395297288895, 0.019012698903679848, 0.07383184134960175, 0.11073657125234604, 0.0368414968252182, -0.0729052945971489, 0.001982470043003559, -0.006313489284366369, -0.039427030831575394, 0.11933320760726929, 0.17963355779647827, -0.11991413682699203, -0.05106910318136215, 0.27167606353759766, 0.0031242913100868464, 0.19481229782104492, -0.01315275114029646, 0.043591804802417755, -0.04484925419092178, 0.04572054371237755, -0.05338600277900696, -0.04086209088563919, 0.2094656229019165, 0.08045925945043564, -0.17165091633796692, -0.08549032360315323, -0.05912299454212189, 0.07081323862075806, 0.10728751868009567, 0.0013539529172703624, -0.04156802222132683, 0.0004610282776411623, 0.0014198932331055403, 0.08339415490627289, -0.14520122110843658, 0.11816094070672989, -0.03172019124031067, 0.05612684786319733, 0.017555562779307365, -0.045326150953769684, 0.04264266416430473, 0.07474290579557419, 0.26618310809135437, 0.0904107540845871, -0.040318213403224945, -0.0892091691493988, -0.12260187417268753, 0.010461576282978058, 0.029102616012096405, -0.03534553572535515, 0.0037547778338193893, -0.020087555050849915, 0.0318896509706974, 0.008264793083071709, 0.016230624169111252, -0.08987458795309067, -0.03175399824976921, -0.027736429125070572, -0.023839212954044342, 0.10733365267515182, -0.09495144337415695, -0.1444292515516281, -0.15713949501514435, 0.04191131144762039, -0.0766405463218689, -0.056593164801597595, -0.054507751017808914, -0.05239389091730118, -0.0311186034232378, -0.03773957118391991, 0.09099467098712921, -0.0021037792321294546, 0.14807306230068207, -0.1920108050107956, -0.04220759496092796, 0.051812779158353806, -0.07607918977737427, -0.08729588985443115, 0.03410962224006653, 0.12136995792388916, 0.05116051807999611, 0.11504370719194412, 0.013609255664050579, 0.09567681699991226, 0.0045484392903745174, -0.06713183224201202, 0.15302421152591705, -0.14069625735282898, -0.27875974774360657, -0.03836318850517273, 0.016946332529187202, 0.1615200787782669, -0.05613167956471443, 0.031766023486852646, 0.3335736393928528, 0.27782970666885376, -0.1428707242012024, 0.25916144251823425, 0.019178593531250954, 0.004398873541504145, -0.19130495190620422, -0.10125631093978882, 0.025324683636426926, 0.04740457236766815, 0.12032642960548401, -0.14564448595046997, -0.010732659138739109, -0.04543145373463631, -0.025908485054969788, 0.10386138409376144, -0.12300799041986465, -0.07263197749853134, 0.07765276730060577, 0.039809420704841614, 0.1808302253484726, 0.03932500258088112, 0.0014799144119024277, 0.13626977801322937, 0.06612244248390198, 0.019124457612633705, 0.05216038227081299, 0.08028066903352737, -0.018944554030895233, 0.14207926392555237, 0.05448179319500923, -0.02551644667983055, 0.052681710571050644, -0.0054580713622272015, -0.03219012916088104, 0.015605825930833817, -0.183198019862175, -0.10147556662559509, -0.0561356320977211, -0.10798973590135574, -0.04978342354297638, 0.056853994727134705, -0.12395523488521576, -0.007896827533841133, -0.03841273859143257, 0.03718273714184761, -0.07831971347332001, -0.09360362589359283, -0.036494381725788116, 0.1351792961359024, 0.07210618257522583, 0.04471297934651375, 0.035655103623867035, -0.07390819489955902, 0.07097936421632767, 0.21671734750270844, 0.08159157633781433, 0.028919655829668045, -0.19545674324035645, -0.024042490869760513, -0.0803457647562027, 0.06306298077106476, -0.08856996893882751, -0.016788700595498085, 0.11923003196716309, 0.08616556972265244, 0.05413002520799637, 0.09640096127986908, -0.045083072036504745, 0.021686913445591927, 0.02684609219431877, -0.15131035447120667, -0.18501274287700653, -0.08534606546163559, -0.03519878163933754, 0.11561143398284912, -0.06398691236972809, 0.10897188633680344, -0.13615410029888153, 0.010051886551082134, -0.006060056854039431, 0.02693452313542366, -0.03596206381917, -0.11251141875982285, 0.15348562598228455, 0.11999429017305374, -0.06767056882381439, 0.03127254918217659, -0.09527092427015305, -0.04423454403877258, 0.12686803936958313, -0.013623855076730251, -0.0371493324637413, -0.054547641426324844, -0.03628576174378395, 0.15247689187526703, -0.03436964750289917, 0.008244883269071579, -0.041229065507650375, -0.18217355012893677, 0.0798322781920433, 0.09045056998729706, 0.019827889278531075, -0.031874191015958786, -0.09797266125679016, -0.010231015272438526, -0.0011165260802954435, 0.11730700731277466, -0.10696814209222794, -0.10933240503072739, -0.15144047141075134, 0.06713984161615372, -0.0007159380475059152, 0.18502596020698547, -0.06394898891448975, -0.08904669433832169, -0.12429379671812057, 0.02344517596065998, -0.0027384376153349876, -0.042264558374881744, 0.01618490368127823, 0.07992301136255264, -0.04095321521162987, 0.02075677551329136, -0.06651144474744797, 0.06372585147619247, -0.11786920577287674, 0.09625071287155151, 0.01063506118953228, 0.016993753612041473, -0.0417880080640316, -0.01618220843374729, 0.039470795542001724, -0.057925306260585785, 0.07921463251113892, 0.011758086271584034, 0.0010938759660348296, 0.10196787863969803, -0.0034960443153977394, 0.06409632414579391, -0.05372481048107147, -0.023290161043405533, 0.06578411161899567, -0.05874887853860855, -0.03370826691389084, -0.1573946475982666, -0.0709633082151413, 0.020051732659339905, -0.04775108024477959, 0.002077929675579071, 0.03673801198601723, 0.062159497290849686, -0.06937079131603241, -0.12125655263662338, -0.043812792748212814, -0.028638383373618126, 0.021301284432411194, 0.10829301923513412, -0.07526551932096481, 0.1547859013080597, -0.052787959575653076, -0.00020603960729204118, 0.07437096536159515, 0.04048224538564682, 0.01393822580575943, -0.10422444343566895, -0.04698587954044342, -0.11035211384296417, 0.1502903699874878, -0.007902312092483044, -0.03533121198415756, 0.03719403222203255, -0.11946307867765427, -0.1572723090648651, 0.03418220207095146, 0.10199101269245148, 0.0448341928422451, 0.025807438418269157, 0.027079269289970398, -0.04042419046163559, -0.021270349621772766, -0.07034418731927872, 0.0882953479886055, -0.12085357308387756, -0.09669415652751923, 0.09555385261774063, 0.12178351730108261, -0.0036850625183433294, -0.07441367954015732, 0.11554073542356491, -0.021787192672491074, 0.05525410920381546, -0.02971339225769043, 0.10308072715997696, 0.0796005055308342, -0.12273547053337097, 0.005693064536899328, -0.036891788244247437, -0.0741485133767128, -0.12975730001926422, 0.019545545801520348, -0.061916105449199677, -0.13383042812347412, 0.12179028987884521, -0.09376577287912369, 0.030037038028240204, -0.10506992787122726, 0.021338803693652153, 0.01864001713693142, 0.061665527522563934, -0.10988292098045349, 0.08575301617383957, 0.13424484431743622, -0.043199893087148666, -0.07184189558029175, -0.12455986440181732, -0.05022053420543671, -0.04231856390833855, -0.13957437872886658, -0.11600435525178909, 0.0100301094353199, -0.023418782278895378, -0.05818291753530502, 0.0015462689334526658, -0.03659068048000336, 0.008594646118581295, 0.021907730028033257, 0.04032021388411522, -0.02693161368370056, 0.05134565755724907, -0.057569269090890884, -0.052510857582092285, 0.11489357799291611, 0.04113486409187317, -0.03561042994260788, -0.052359987050294876, 0.12997733056545258, -0.11959461867809296, 0.07662346214056015, -0.020313527435064316, 0.017129231244325638, -0.06435854732990265, 0.17131924629211426, 0.11673715710639954, -0.1367570012807846, -0.005008010193705559, -0.08210669457912445, 0.020409544929862022, 0.023555370047688484, 0.13693512976169586, -0.03411718085408211, -0.0012358218664303422, -0.1580323874950409, 0.018575575202703476, -0.18557456135749817, -0.03716109320521355, 0.04671547934412956, 0.09917585551738739, 0.15293832123279572, -0.0034432117827236652, -0.1263325810432434, 0.10424192249774933, -0.2118520885705948, 0.0907607227563858, 0.05121984705328941, -0.11874113976955414, -0.06765396893024445, -0.06795281916856766, 0.1198519766330719, 0.009196433238685131, 0.2040700763463974, -0.013615905307233334, -0.09132910519838333, -0.07060808688402176, -0.01980910450220108, -0.030524181202054024, 0.09714830666780472, 0.041414931416511536, 0.04653804749250412, 0.12821412086486816, 0.00368314771912992, 0.07533777505159378, 0.060310911387205124, 0.02759413793683052, -0.012300663627684116, 0.04076618701219559, 0.08261215686798096, -0.14588621258735657, -0.1659701019525528, 0.1326720416545868, 0.025149408727884293, 0.11792458593845367, 0.03658788278698921, -0.1549617499113083, 0.06687124073505402, 0.2523096203804016, -0.11147607117891312, 0.02505038119852543, 0.12737524509429932, -0.0366884209215641, 0.0672016367316246, 0.1144871786236763, -0.02633814327418804, -0.05217865854501724, -0.011363590136170387, 0.10233135521411896, 0.028660254552960396, -0.04646271467208862, -0.02340836264193058, -0.03373933956027031, -0.019070526584982872, -0.011738128960132599, -0.0909019410610199, -0.1543993502855301, -0.10471053421497345, -0.16619662940502167, 0.04399140924215317, -0.04626438021659851, 0.13418889045715332, 0.09469578415155411, -0.012723101302981377, 0.04568437114357948, 0.028575526550412178, 0.07275456190109253, 0.07916246354579926, -0.02939477376639843, -0.036159269511699677 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # RerankerModel_chat_bot This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 3 ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.1
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "bert-base-uncased", "model-index": [{"name": "RerankerModel_chat_bot", "results": []}]}
text-classification
Shakhovak/RerankerModel_chat_bot
[ "transformers", "safetensors", "bert", "text-classification", "generated_from_trainer", "base_model:bert-base-uncased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "has_space", "region:us" ]
2024-02-10T13:13:53+00:00
[]
[]
TAGS #transformers #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us
# RerankerModel_chat_bot This model is a fine-tuned version of bert-base-uncased on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_ratio: 0.1 - num_epochs: 3 ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.1
[ "# RerankerModel_chat_bot\n\nThis model is a fine-tuned version of bert-base-uncased on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_ratio: 0.1\n- num_epochs: 3", "### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n", "# RerankerModel_chat_bot\n\nThis model is a fine-tuned version of bert-base-uncased on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_ratio: 0.1\n- num_epochs: 3", "### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1" ]
[ 68, 35, 6, 12, 8, 3, 106, 33 ]
[ "passage: TAGS\n#transformers #safetensors #bert #text-classification #generated_from_trainer #base_model-bert-base-uncased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #has_space #region-us \n# RerankerModel_chat_bot\n\nThis model is a fine-tuned version of bert-base-uncased on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_ratio: 0.1\n- num_epochs: 3### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.2.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1" ]
[ -0.07152149826288223, 0.08271459490060806, -0.002031854121014476, 0.08368208259344101, 0.1286334991455078, 0.029067400842905045, 0.12398061901330948, 0.11364853382110596, -0.07629209756851196, 0.06351219117641449, 0.05717819929122925, 0.03799600154161453, 0.03334668651223183, 0.15386663377285004, -0.05986856296658516, -0.19436262547969818, 0.037888094782829285, -0.042056605219841, -0.042392998933792114, 0.07991084456443787, 0.08521056920289993, -0.11993788182735443, 0.09577514231204987, -0.007942375726997852, -0.15356603264808655, 0.0019527666736394167, -0.019096853211522102, -0.042764436453580856, 0.0992855653166771, -0.015049837529659271, 0.12408143281936646, 0.04178360849618912, 0.15031659603118896, -0.1961987018585205, 0.003142318921163678, 0.08567477762699127, 0.030692696571350098, 0.07683588564395905, 0.02075064182281494, 0.023614222183823586, 0.0647580698132515, -0.14061619341373444, 0.11718165874481201, 0.024278895929455757, -0.08439677953720093, -0.058398161083459854, -0.08770725131034851, 0.010842646472156048, 0.07817517966032028, 0.08411622047424316, 0.017885886132717133, 0.1621498465538025, -0.11343926936388016, 0.06738182157278061, 0.1936621367931366, -0.2867938280105591, -0.06405328214168549, 0.0918072760105133, 0.056375037878751755, 0.08696559816598892, -0.1302659660577774, 0.0033734848257154226, 0.03194814175367355, 0.02090698480606079, 0.11587006598711014, -0.032406676560640335, -0.14763085544109344, -0.02768399938941002, -0.11169910430908203, -0.008881689049303532, 0.11383809894323349, 0.04086805880069733, -0.05576784163713455, -0.05905766785144806, -0.0663839802145958, -0.08204052597284317, -0.029093824326992035, -0.05330017954111099, 0.056067321449518204, -0.04381750524044037, -0.052740611135959625, -0.05763392150402069, -0.08706220984458923, -0.07642170041799545, -0.005921896081417799, 0.1205410435795784, 0.016686875373125076, 0.018121041357517242, -0.025805577635765076, 0.10557891428470612, -0.027288198471069336, -0.1243056133389473, -0.013966167345643044, -0.0026747779920697212, -0.10209996998310089, -0.07888196408748627, -0.04851910471916199, -0.04274917021393776, 0.02202056534588337, 0.15827113389968872, -0.07729443162679672, 0.07141935080289841, -0.027187611907720566, 0.0021124258637428284, -0.03437883406877518, 0.14698652923107147, -0.02053760550916195, -0.050089675933122635, 0.02605614624917507, 0.10581089556217194, 0.025776609778404236, -0.01631809026002884, -0.09095346927642822, -0.026960089802742004, 0.09603934735059738, 0.047521915286779404, -0.0659240186214447, 0.049689583480358124, -0.03625141456723213, -0.015529504977166653, 0.03680984303355217, -0.10958577692508698, 0.052175041288137436, 0.013873368501663208, -0.06969094276428223, -0.06266817450523376, 0.0654156282544136, 0.013323678635060787, 0.0001267174375243485, 0.0850803330540657, -0.08828538656234741, -0.001408479642122984, -0.08396930247545242, -0.08178725838661194, 0.013618712313473225, -0.11916279792785645, -0.01456714142113924, -0.09466788917779922, -0.18358393013477325, -0.03527362644672394, 0.042370688170194626, -0.05100693926215172, -0.021369481459259987, -0.06677823513746262, -0.07227171212434769, 0.020085690543055534, -0.006919071543961763, 0.1262856274843216, -0.06370334327220917, 0.04401860386133194, -0.04262969642877579, 0.03305981308221817, -0.024576224386692047, 0.03549479320645332, -0.1048661544919014, 0.015478688292205334, -0.1934228092432022, 0.051350340247154236, -0.08321273326873779, 0.03716198727488518, -0.09501617401838303, -0.06575814634561539, 0.005517289508134127, -0.03215687721967697, 0.050648123025894165, 0.13694064319133759, -0.1684938669204712, -0.0054207900539040565, 0.12163276225328445, -0.09076805412769318, -0.048249971121549606, 0.08150159567594528, -0.06080654263496399, 0.044198859483003616, 0.06551837921142578, 0.1503988355398178, 0.07479363679885864, -0.16147947311401367, -0.012741528451442719, 0.017757859081029892, 0.044422462582588196, 0.024886354804039, 0.027300652116537094, -0.002482728101313114, 0.02359316311776638, 0.006161567289382219, -0.04377109557390213, -0.013176745735108852, -0.06899816542863846, -0.08051151782274246, -0.05869915708899498, -0.09786881506443024, 0.032806746661663055, 0.01601623184978962, 0.04502461850643158, -0.07011078298091888, -0.1020488291978836, 0.08409035205841064, 0.12157949805259705, -0.05253509059548378, 0.022090043872594833, -0.07283920794725418, 0.004614419303834438, 0.01030732225626707, -0.024354707449674606, -0.1819443702697754, -0.13389162719249725, 0.0571502149105072, -0.03392699360847473, 0.05768528953194618, 0.05569654703140259, 0.07664424180984497, 0.07345122843980789, -0.045511193573474884, -0.011391323059797287, -0.06857506185770035, 0.0193803533911705, -0.12218207865953445, -0.21924981474876404, -0.028661780059337616, -0.04643506556749344, 0.20054273307323456, -0.20986266434192657, 0.013995382003486156, -0.012655173428356647, 0.13933274149894714, 0.04454571753740311, -0.04373219609260559, -0.0018471903167665005, 0.04070170968770981, -0.008342575281858444, -0.09018968790769577, 0.02832670696079731, 0.007007925771176815, -0.07530078291893005, -0.04291734844446182, -0.17089103162288666, -0.0032556517980992794, 0.083781398832798, 0.056407053023576736, -0.10099285840988159, -0.02255718968808651, -0.03736988827586174, -0.041027914732694626, -0.06758657842874527, 0.005844754166901112, 0.18995197117328644, 0.01276041567325592, 0.1286935955286026, -0.05075743794441223, -0.04094550013542175, -0.0029385769739747047, 0.010496257804334164, 0.020024187862873077, 0.046760059893131256, 0.0738472044467926, -0.1322929412126541, 0.07924846559762955, 0.125132218003273, -0.09627328813076019, 0.11801499873399734, -0.026881735771894455, -0.0604359395802021, -0.007102270144969225, -0.0019666843581944704, -0.010994850657880306, 0.12105925381183624, -0.07501883804798126, 0.01634668931365013, 0.019379885867238045, 0.01815229468047619, 0.005550697445869446, -0.15592747926712036, 0.012304889038205147, 0.004159208852797747, -0.04726307839155197, 0.003014987800270319, -0.033268678933382034, 0.03916897252202034, 0.08758777379989624, 0.01123750302940607, -0.07675467431545258, 0.018783338367938995, -0.006512564606964588, -0.06003834307193756, 0.18766848742961884, -0.1272939294576645, -0.10581088811159134, -0.0831596627831459, 0.06800680607557297, -0.05092092975974083, -0.006636971607804298, 0.016066696494817734, -0.0685102641582489, -0.06530396640300751, -0.10358614474534988, -0.054994817823171616, 0.023989012464880943, 0.014336257241666317, 0.02352317050099373, 0.0016989544965326786, 0.07831922173500061, -0.1326267570257187, -0.014051166363060474, -0.026205776259303093, -0.1299103945493698, 0.021289058029651642, 0.052945394068956375, 0.08065970987081528, 0.12081870436668396, -0.03508454188704491, -0.0060267201624810696, -0.02331211417913437, 0.18669629096984863, -0.062425289303064346, 0.01513705961406231, 0.07896482199430466, -0.0042693899013102055, 0.04230194538831711, 0.14935266971588135, 0.025803998112678528, -0.0908118486404419, 0.034243904054164886, 0.07312501966953278, -0.01907598040997982, -0.22272078692913055, -0.05101051554083824, -0.017031965777277946, -0.060398004949092865, 0.09029944241046906, 0.04326368123292923, 0.02157943695783615, 0.045792967081069946, -0.002969691064208746, 0.018031427636742592, 0.012037073262035847, 0.07151660323143005, 0.0755358338356018, 0.04963874816894531, 0.10567425936460495, -0.01277972012758255, -0.022290309891104698, 0.054058995097875595, -0.004303157329559326, 0.22209994494915009, -0.011971589177846909, 0.0898565724492073, 0.01847941428422928, 0.12448178231716156, -0.027699818834662437, 0.05555842071771622, 0.03801485151052475, -0.015445798635482788, -0.016198936849832535, -0.06377467513084412, -0.03975161164999008, 0.04278728738427162, -0.028541119769215584, 0.031373001635074615, -0.0881701186299324, 0.06372472643852234, 0.03492766246199608, 0.280459463596344, 0.0405682772397995, -0.29481664299964905, -0.060401998460292816, 0.00639347406104207, -0.0274899210780859, -0.049172841012477875, -0.0019576973281800747, 0.14758123457431793, -0.12606438994407654, 0.054563093930482864, -0.0438377670943737, 0.0812639370560646, -0.019664397463202477, 0.021790649741888046, 0.016593366861343384, 0.13716480135917664, -0.010794861242175102, 0.05321724712848663, -0.2187010794878006, 0.20836368203163147, 0.03392687812447548, 0.1161428913474083, -0.05980292707681656, 0.013648059219121933, 0.03225114196538925, 0.12661994993686676, 0.0935165286064148, -0.01482507586479187, -0.04646870121359825, -0.1638089269399643, -0.10554612427949905, 0.05122632905840874, 0.10765990614891052, -0.01734205149114132, 0.08148980140686035, -0.04007573425769806, -0.003872103523463011, 0.05792353302240372, -0.05317848175764084, -0.18720215559005737, -0.09175765514373779, -0.0002602596941869706, 0.054467953741550446, -0.03492089733481407, -0.06458228826522827, -0.08875622600317001, -0.030453519895672798, 0.1640482395887375, 0.028383465483784676, -0.014483324252068996, -0.1316734403371811, 0.0414506234228611, 0.10693427175283432, -0.05188525468111038, 0.021600838750600815, -0.002668060129508376, 0.10208253562450409, 0.018316026777029037, -0.07958471775054932, 0.0632253959774971, -0.08480055630207062, -0.1677180975675583, -0.06476977467536926, 0.0877237617969513, 0.08166749775409698, 0.05368553847074509, 0.03519219532608986, 0.02609599195420742, 0.01717928797006607, -0.08017819374799728, -0.004963782150298357, 0.08651383966207504, 0.08532395958900452, 0.056271743029356, -0.10428567230701447, -0.012630623765289783, -0.02617610991001129, 0.009818683378398418, 0.11287957429885864, 0.2244172990322113, -0.09244456142187119, 0.0935957282781601, 0.10467816889286041, -0.07638832926750183, -0.20320728421211243, 0.0689973533153534, 0.08977577835321426, 0.017011228948831558, 0.04795118048787117, -0.171944260597229, 0.13701295852661133, 0.10128810256719589, -0.034554820507764816, 0.01262986846268177, -0.2646208703517914, -0.13546600937843323, 0.11457505077123642, 0.07983420789241791, 0.0739188864827156, -0.12128622084856033, -0.012632442638278008, -0.03751545026898384, -0.08726180344820023, 0.14130465686321259, -0.17219242453575134, 0.12256435304880142, 0.007141442038118839, 0.09016484022140503, 0.028925461694598198, -0.03550131246447563, 0.12966322898864746, 0.03987491503357887, 0.10738734155893326, -0.060680948197841644, 0.006563510280102491, 0.06029916927218437, -0.08184255659580231, 0.05111046880483627, -0.07223894447088242, 0.05402297154068947, -0.09312282502651215, 0.0090257553383708, -0.07475247979164124, 0.09447367489337921, -0.03292261064052582, -0.06168189272284508, -0.04282403737306595, 0.04962872341275215, 0.053299009799957275, -0.019854070618748665, 0.08602773398160934, 0.010202287696301937, 0.11064989864826202, 0.1688278317451477, 0.11823751777410507, -0.06108127161860466, -0.020447880029678345, 0.030819710344076157, -0.022797243669629097, 0.07660689204931259, -0.07541581988334656, 0.03710678592324257, 0.11274208128452301, 0.025775868445634842, 0.13342726230621338, 0.0315268449485302, -0.04497421532869339, -0.006617777980864048, 0.043253909796476364, -0.14550939202308655, -0.09687919169664383, 0.03755737841129303, 0.006249791942536831, -0.12209208309650421, 0.023881640285253525, 0.1467381864786148, -0.0665963888168335, 0.0004513147578109056, -0.012559073977172375, 0.01924939826130867, -0.033456236124038696, 0.184322789311409, 0.02803133800625801, 0.0467691533267498, -0.08974216133356094, 0.1360405683517456, 0.08218805491924286, -0.06262737512588501, 0.05806531012058258, 0.0597468838095665, -0.1162545382976532, -0.03220954164862633, 0.10419170558452606, 0.18522831797599792, -0.0398237518966198, -0.030583802610635757, -0.08353268355131149, -0.08466797322034836, 0.021563822403550148, 0.16495577991008759, 0.04767708107829094, 0.0045638796873390675, -0.029122596606612206, 0.040397725999355316, -0.12148603051900864, 0.06374068558216095, 0.013041388243436813, 0.07252415269613266, -0.12483642250299454, 0.13143989443778992, -0.006130356341600418, 0.009730499237775803, -0.026807552203536034, 0.03472623601555824, -0.10643958300352097, -0.0418037474155426, -0.11272699385881424, 0.010483241640031338, -0.0033483575098216534, 0.007223170716315508, 0.008546233177185059, -0.05392191559076309, -0.036694034934043884, 0.02992592751979828, -0.07973140478134155, -0.04321281239390373, 0.008306406438350677, 0.06151493266224861, -0.1511661559343338, -0.032259855419397354, 0.022590359672904015, -0.08354955166578293, 0.07743983715772629, 0.06443009525537491, 0.01684107445180416, 0.03831826522946358, -0.19231067597866058, -0.03486024960875511, 0.043475788086652756, 0.026863932609558105, 0.07737518101930618, -0.057660430669784546, -0.03131833299994469, -0.028039999306201935, 0.061349764466285706, 0.022981923073530197, 0.10862376540899277, -0.09790544956922531, 0.0013896074378862977, -0.0360102653503418, -0.06622502952814102, -0.05891367420554161, 0.03366091102361679, 0.11980624496936798, 0.01646493189036846, 0.17084646224975586, -0.0885409340262413, 0.022876212373375893, -0.17340624332427979, -0.017827443778514862, 0.0017955341609194875, -0.03906022757291794, -0.09851301461458206, -0.04618331044912338, 0.08704317361116409, -0.04808336868882179, 0.07051441073417664, 0.007934394292533398, 0.09689123183488846, 0.03617853671312332, -0.046893663704395294, -0.048292696475982666, 0.021823346614837646, 0.129975825548172, 0.08108682930469513, -0.01619093306362629, 0.09338754415512085, -0.016271498054265976, 0.08212558925151825, 0.020792648196220398, 0.23782816529273987, 0.1418648660182953, -0.05077466741204262, 0.08791067451238632, 0.07318345457315445, -0.12930898368358612, -0.11535990238189697, 0.12285209447145462, -0.06014230474829674, 0.11476154625415802, -0.08057080209255219, 0.18524757027626038, 0.06538690626621246, -0.16455800831317902, 0.028564682230353355, -0.07218242436647415, -0.10308551043272018, -0.1311565488576889, 0.0003783780266530812, -0.09659534692764282, -0.12619620561599731, 0.010657910257577896, -0.1066160798072815, 0.052559856325387955, 0.08534304052591324, 0.0043089063838124275, 0.025423763319849968, 0.17043277621269226, -0.07509846240282059, 0.029872115701436996, 0.04249226301908493, 0.016728417947888374, -0.014609787613153458, -0.06829627603292465, -0.09745530039072037, 0.03927793353796005, 0.03923400864005089, 0.053608305752277374, -0.051137618720531464, 0.034624893218278885, 0.044113364070653915, -0.00023036937636788934, -0.07173019647598267, 0.03398311138153076, 0.00834308099001646, 0.031675923615694046, 0.040848035365343094, 0.04551830142736435, 0.009314443916082382, -0.020794861018657684, 0.2758263647556305, -0.08321264386177063, -0.08311165869235992, -0.13255146145820618, 0.22014302015304565, 0.025808651000261307, 0.00486266054213047, 0.04174390807747841, -0.09849587827920914, -0.02425296977162361, 0.14029359817504883, 0.16136839985847473, -0.10451650619506836, 0.00030998309375718236, -0.04676942899823189, -0.015361280180513859, -0.0605628564953804, 0.11258232593536377, 0.07662002742290497, -0.029050355777144432, -0.06691905856132507, 0.014764560386538506, 0.024572066962718964, -0.030080782249569893, -0.09308955818414688, 0.010393643751740456, -0.013776799663901329, 0.015502399764955044, -0.032880403101444244, 0.0379507839679718, 0.01474465150386095, -0.20115885138511658, 0.03168603777885437, -0.1775321513414383, -0.1467142552137375, -0.029740430414676666, 0.07357297092676163, -0.026782944798469543, 0.05079518258571625, -0.03228573501110077, -0.014896840788424015, 0.10291767120361328, -0.04455579072237015, -0.03325602412223816, -0.12792032957077026, 0.08852090686559677, -0.07470555603504181, 0.24607020616531372, -0.005950085818767548, 0.06759598106145859, 0.09217189997434616, 0.01959925703704357, -0.11566002666950226, 0.04281178489327431, 0.05632234737277031, -0.10901376605033875, 0.013281093910336494, 0.13850684463977814, -0.07432734221220016, 0.11862989515066147, 0.03453940153121948, -0.13383828103542328, -0.004248746670782566, -0.04656875133514404, -0.040440745651721954, -0.057762615382671356, -0.021665941923856735, -0.057224564254283905, 0.1663902848958969, 0.18990422785282135, -0.008200562559068203, 0.019072595983743668, -0.06817596405744553, 0.028563035652041435, 0.05666632950305939, 0.03348013013601303, -0.06240564584732056, -0.20681791007518768, 0.04716327041387558, 0.031102050095796585, 0.026647185906767845, -0.20000188052654266, -0.08911807835102081, 0.02192857302725315, -0.020282255485653877, -0.06932635605335236, 0.09364546090364456, 0.0414138101041317, 0.02177293226122856, -0.03199263662099838, -0.15297362208366394, -0.032666511833667755, 0.1383461356163025, -0.16065272688865662, -0.06726212054491043 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
null
tyemel/vit-base-patch16-224-in21k-finetuned-lora-food101
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
2024-02-10T13:16:41+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 31, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.06646376848220825, 0.2168014943599701, -0.00225935154594481, 0.023818302899599075, 0.1271018385887146, -0.001635765191167593, 0.04218708351254463, 0.13324736058712006, -0.020175931975245476, 0.11144465953111649, 0.046588581055402756, 0.09377603232860565, 0.09928803145885468, 0.18404334783554077, 0.04859916493296623, -0.2059975117444992, 0.007056170143187046, -0.09090408682823181, 0.014076028019189835, 0.1116579994559288, 0.13719257712364197, -0.10291384905576706, 0.08272874355316162, -0.04045208916068077, -0.02019004337489605, 0.00012576708104461432, -0.09259183704853058, -0.07032395154237747, 0.06885425746440887, 0.06264153122901917, 0.051234472543001175, 0.001456156256608665, 0.09140396863222122, -0.2864592671394348, 0.017265573143959045, 0.08406311273574829, 0.0027674848679453135, 0.06290827691555023, 0.07236549258232117, -0.07389893382787704, 0.11328595131635666, -0.08021481335163116, 0.13019037246704102, 0.08625296503305435, -0.062064990401268005, -0.23071379959583282, -0.07525765895843506, 0.0963398814201355, 0.12251301854848862, 0.06215599179267883, -0.022921854630112648, 0.15455181896686554, -0.06248689442873001, 0.012971068732440472, 0.1294165402650833, -0.11526761949062347, -0.05572471022605896, 0.061741601675748825, 0.11775490641593933, 0.10740239918231964, -0.14110268652439117, -0.0017287094378843904, 0.04900608956813812, 0.029121357947587967, 0.08589313924312592, 0.022661056369543076, 0.12003941088914871, 0.04652795568108559, -0.13695219159126282, -0.04037507623434067, 0.12011898308992386, 0.038862764835357666, -0.06446044892072678, -0.2168138176202774, -0.006778308190405369, -0.0601806715130806, -0.014732478186488152, -0.07019448280334473, 0.039128515869379044, -0.02470310963690281, 0.07317749410867691, -0.04465159401297569, -0.1063927412033081, -0.0421026237308979, 0.0892222449183464, 0.07748593389987946, 0.011527054943144321, -0.02519804798066616, 0.04627908393740654, 0.13455867767333984, 0.05402068421244621, -0.10399353504180908, -0.07017925381660461, -0.06942764669656754, -0.09420394152402878, -0.04035796597599983, 0.056760527193546295, 0.031942449510097504, 0.02665667235851288, 0.22703726589679718, 0.016653569415211678, 0.04155244305729866, 0.0224777739495039, 0.01032855175435543, 0.043662428855895996, 0.0955500528216362, -0.05303520709276199, -0.15660029649734497, -0.04072032496333122, 0.09077946096658707, -0.0027527001220732927, -0.036689214408397675, -0.03966725245118141, 0.03849169611930847, 0.06843466311693192, 0.13122352957725525, 0.07552056759595871, -0.017929591238498688, -0.04813180863857269, -0.030096933245658875, 0.23523783683776855, -0.1493375599384308, 0.04426715523004532, -0.02271856553852558, -0.01804111897945404, -0.03908449783921242, 0.03597262129187584, 0.022118929773569107, -0.000004518366949923802, 0.09706240892410278, -0.058981191366910934, -0.05378659814596176, -0.10168042778968811, -0.03272576630115509, 0.04088849574327469, -0.013975566253066063, -0.010589460842311382, -0.09025166928768158, -0.09490354359149933, -0.04766594246029854, 0.05537205561995506, -0.05123869329690933, -0.03770573064684868, 0.009465423412621021, -0.08151785284280777, -0.005444355774670839, -0.005417742300778627, 0.10699385404586792, -0.03222226724028587, 0.04445803165435791, -0.027600755915045738, 0.05225523188710213, 0.09919606149196625, 0.031576547771692276, -0.0773419588804245, 0.0561848059296608, -0.22559374570846558, 0.07503069192171097, -0.11481974273920059, 0.04335082694888115, -0.1704932004213333, -0.042439818382263184, 0.005444696638733149, 0.0139949731528759, 0.013206101022660732, 0.12720820307731628, -0.19255615770816803, -0.01654396951198578, 0.13260798156261444, -0.09212633967399597, -0.118110790848732, 0.07884611934423447, -0.029701577499508858, 0.1624738723039627, 0.04682036489248276, -0.027025915682315826, 0.09224298596382141, -0.16434773802757263, -0.07092688232660294, -0.00949116237461567, -0.01727987825870514, 0.12109188735485077, 0.07512219995260239, -0.05991523340344429, 0.046571120619773865, 0.02832140028476715, -0.038078423589468, -0.04424772411584854, -0.050857074558734894, -0.10884185880422592, -0.01070026308298111, -0.08987759798765182, 0.04065500199794769, -0.01250192429870367, -0.07916021347045898, -0.029885273426771164, -0.18612512946128845, -0.0030564051121473312, 0.10038342326879501, 0.0035033065360039473, -0.005652366206049919, -0.08666291832923889, 0.026358824223279953, -0.03112892620265484, -0.008404186926782131, -0.16764774918556213, -0.04399421438574791, 0.046902090311050415, -0.16094985604286194, 0.020117372274398804, -0.06413903087377548, 0.06334125250577927, 0.03641495108604431, -0.05590536445379257, -0.0248766727745533, -0.01730942726135254, 0.011945613659918308, -0.05083848536014557, -0.18994836509227753, -0.056277405470609665, -0.037882111966609955, 0.149809330701828, -0.25956398248672485, 0.032966937869787216, 0.051140617579221725, 0.14649195969104767, 0.00406361510977149, -0.05115427449345589, 0.01429014839231968, -0.05360214412212372, -0.054652128368616104, -0.06746816635131836, -0.006135428790003061, -0.027576493099331856, -0.05147203803062439, 0.019243421033024788, -0.1755700707435608, -0.021410830318927765, 0.09424154460430145, 0.12876708805561066, -0.1486445665359497, -0.018640631809830666, -0.048725154250860214, -0.06339836865663528, -0.0715010017156601, -0.07038594037294388, 0.10712739825248718, 0.0513901449739933, 0.04796046018600464, -0.07435787469148636, -0.07092321664094925, 0.02726263552904129, 0.006906150374561548, -0.03382374346256256, 0.08727246522903442, 0.05199531093239784, -0.09209315478801727, 0.0756213590502739, 0.1092359870672226, 0.07177663594484329, 0.09363535046577454, 0.01574566215276718, -0.11756632477045059, -0.028492970392107964, 0.036266472190618515, 0.02740776725113392, 0.1465986967086792, -0.05952361226081848, 0.04016614332795143, 0.04494241625070572, -0.04170418903231621, 0.022319864481687546, -0.08787637203931808, 0.024075502529740334, 0.025203049182891846, -0.0034381982404738665, 0.06284574419260025, -0.02525499276816845, -0.0050758360885083675, 0.07016654312610626, 0.047779910266399384, 0.04621000960469246, 0.009655474685132504, -0.01720241829752922, -0.1047825813293457, 0.16950392723083496, -0.0951867327094078, -0.269941508769989, -0.17632324993610382, 0.026197833940386772, 0.04035249724984169, -0.022378476336598396, 0.031619444489479065, -0.07056326419115067, -0.10630585998296738, -0.1060405746102333, -0.002429972169920802, 0.01714223250746727, -0.06364088505506516, -0.0741225928068161, 0.07348573952913284, 0.04382912442088127, -0.14902326464653015, 0.038552410900592804, 0.055694397538900375, -0.057955220341682434, -0.0233661737293005, 0.09118817001581192, 0.12397737801074982, 0.14583967626094818, -0.021366750821471214, -0.028626007959246635, 0.029004426673054695, 0.19620531797409058, -0.13469526171684265, 0.10371150821447372, 0.13814030587673187, -0.04545360431075096, 0.08360563963651657, 0.1560150384902954, 0.029186224564909935, -0.08317049592733383, 0.05044832453131676, 0.04082648828625679, -0.043159641325473785, -0.2666129767894745, -0.0534592866897583, 0.012832709588110447, -0.06255637854337692, 0.09786593168973923, 0.10183793306350708, 0.11542957276105881, 0.034910861402750015, -0.07166364789009094, -0.043925940990448, -0.0058974819257855415, 0.11737963557243347, -0.05490213260054588, -0.012639665976166725, 0.07686592638492584, -0.05086168646812439, 0.005355054512619972, 0.10266812145709991, 0.02973790094256401, 0.17442677915096283, 0.020399179309606552, 0.11231429129838943, 0.06195578724145889, 0.08633565157651901, 0.0007386076031252742, 0.02951662428677082, 0.05147615820169449, 0.017203815281391144, -0.002300140680745244, -0.10421168059110641, -0.006156572140753269, 0.1449710875749588, 0.028103826567530632, 0.029669636860489845, -0.0018948549404740334, -0.005003341939300299, 0.05121048167347908, 0.1746254414319992, -0.011592294089496136, -0.22072425484657288, -0.0845772922039032, 0.06936841458082199, -0.06218599155545235, -0.12968985736370087, -0.026130788028240204, 0.045467354357242584, -0.17519839107990265, 0.026703642681241035, -0.027433741837739944, 0.0919293761253357, -0.09345759451389313, -0.02221956104040146, 0.03687324374914169, 0.084866963326931, -0.014529162086546421, 0.08703910559415817, -0.14498743414878845, 0.11886418610811234, 0.02978132851421833, 0.09024628251791, -0.11081171780824661, 0.07909037172794342, -0.007550720125436783, 0.009180475026369095, 0.19379350543022156, -0.011335089802742004, -0.03514958545565605, -0.08774717897176743, -0.11210042238235474, -0.013537433929741383, 0.12687496840953827, -0.1243172138929367, 0.08773399889469147, -0.015198243781924248, -0.044079482555389404, 0.00937260314822197, -0.12100647389888763, -0.17273177206516266, -0.19628387689590454, 0.05585884302854538, -0.09575839340686798, 0.025643249973654747, -0.11914430558681488, -0.07089093327522278, -0.02952558360993862, 0.241120383143425, -0.1745356321334839, -0.06510113179683685, -0.1468164622783661, -0.046294767409563065, 0.1662203073501587, -0.04437198117375374, 0.0718095526099205, -0.0208172257989645, 0.20345525443553925, 0.005988610442727804, -0.004939318168908358, 0.06724198162555695, -0.08892562240362167, -0.16873881220817566, -0.06771010160446167, 0.1510489284992218, 0.11680185794830322, 0.04907919466495514, -0.002248800592496991, 0.0011772146681323647, -0.016943959519267082, -0.1137804463505745, -0.0033210667315870523, 0.16037839651107788, 0.03878779336810112, 0.025986969470977783, -0.05243593826889992, -0.08797456324100494, -0.06899320334196091, -0.06853509694337845, 0.06221301481127739, 0.19590823352336884, -0.10376439243555069, 0.1700313836336136, 0.147536963224411, -0.07305635511875153, -0.23175598680973053, 0.035342130810022354, 0.04983805492520332, 0.0014306638622656465, 0.04886869341135025, -0.18252557516098022, 0.10521943867206573, 0.019543392583727837, -0.05505957826972008, 0.13485197722911835, -0.1557481735944748, -0.1552847921848297, 0.0722852572798729, 0.03904085233807564, -0.22423844039440155, -0.1354004591703415, -0.09622503817081451, -0.05825018882751465, -0.14065024256706238, 0.06054598465561867, -0.002136280992999673, 0.015948504209518433, 0.03500790148973465, -0.0015643214574083686, 0.027123261243104935, -0.058935679495334625, 0.18609118461608887, -0.004065449349582195, 0.020676052197813988, -0.060264769941568375, -0.0478842556476593, 0.09839435666799545, -0.06130504235625267, 0.12208222597837448, 0.004057085141539574, 0.01594383642077446, -0.10362856835126877, -0.048314861953258514, -0.04328322783112526, 0.05154227837920189, -0.07548051327466965, -0.10070807486772537, -0.043625857681035995, 0.08841723203659058, 0.07005169242620468, -0.03383097052574158, 0.00549331633374095, -0.07189501076936722, 0.10019614547491074, 0.17795267701148987, 0.17573626339435577, 0.009926567785441875, -0.07241068035364151, 0.01677953451871872, -0.04142116755247116, 0.044231921434402466, -0.2513144314289093, 0.03756171092391014, 0.06098250672221184, 0.029438555240631104, 0.09217222779989243, -0.020435843616724014, -0.1820858269929886, -0.04050002992153168, 0.08094815909862518, -0.05452597141265869, -0.22617179155349731, -0.019085140898823738, 0.0954197570681572, -0.2020406424999237, -0.007372708059847355, 0.03995226323604584, -0.048725228756666183, -0.023169852793216705, 0.00010950004070764408, 0.06317184865474701, 0.002471912419423461, 0.09773622453212738, 0.0735151618719101, 0.09715340286493301, -0.08337292820215225, 0.10562895983457565, 0.10150538384914398, -0.09572599828243256, 0.03605884686112404, 0.06754924356937408, -0.05300498008728027, -0.043293699622154236, 0.03665391728281975, 0.033023297786712646, 0.005234600510448217, -0.060321882367134094, 0.013913018628954887, -0.036497246474027634, 0.044923391193151474, 0.08326134830713272, 0.03754979372024536, -0.013354414142668247, 0.06462216377258301, 0.03401726484298706, -0.10898099094629288, 0.10366570204496384, 0.01731540448963642, 0.04105307161808014, -0.08384523540735245, -0.019968897104263306, 0.035425446927547455, 0.030576206743717194, -0.01765924133360386, -0.02306121215224266, -0.02860277332365513, -0.01614218018949032, -0.14299540221691132, -0.023106401786208153, -0.07243485748767853, 0.006181265693157911, 0.014656842686235905, -0.031884219497442245, -0.011233693920075893, 0.02475680410861969, -0.06979699432849884, -0.07426341623067856, -0.006949664559215307, 0.09833318740129471, -0.15115703642368317, 0.008848577737808228, 0.06907843053340912, -0.11088496446609497, 0.08190931379795074, -0.008411259390413761, 0.016245156526565552, 0.022527478635311127, -0.15448406338691711, 0.05601610988378525, 0.0008648968650959432, 0.01916889287531376, 0.025886621326208115, -0.16471809148788452, 0.004104440100491047, -0.04661374166607857, -0.02149827405810356, -0.00004464812809601426, -0.02647159807384014, -0.12325995415449142, 0.06858719140291214, -0.015622655861079693, -0.035931166261434555, -0.02701525390148163, 0.0539589487016201, 0.07888586074113846, -0.027474910020828247, 0.10445091128349304, -0.008690856397151947, 0.04941811040043831, -0.16801609098911285, -0.02470702864229679, -0.04982255399227142, 0.019377702847123146, 0.009884213097393513, -0.007693959400057793, 0.04183054715394974, -0.00976533442735672, 0.21883612871170044, -0.05075952783226967, 0.1607085019350052, 0.05847611650824547, -0.017352959141135216, -0.0007513365126214921, 0.06180921941995621, 0.05997028574347496, 0.04658793285489082, 0.009480604901909828, 0.023740366101264954, -0.022450892254710197, -0.006695089396089315, -0.15932634472846985, 0.01890849508345127, 0.14999441802501678, 0.06301083415746689, 0.024745315313339233, 0.05866100639104843, -0.12775006890296936, -0.12135478109121323, 0.09311001747846603, -0.026755332946777344, 0.00928465835750103, -0.08245618641376495, 0.1358020007610321, 0.14980104565620422, -0.14000412821769714, 0.05256148427724838, -0.06134212389588356, -0.05217423290014267, -0.10388828068971634, -0.12032219022512436, -0.05887215584516525, -0.053666237741708755, 0.002330566756427288, -0.03760887682437897, 0.054546963423490524, 0.03344334661960602, -0.009351172484457493, -0.00022941511997487396, 0.13597318530082703, -0.019751882180571556, -0.0028988157864660025, 0.048313532024621964, 0.03693558648228645, 0.02373051457107067, -0.05275435373187065, 0.02940409444272518, 0.02539868652820587, 0.032232340425252914, 0.06546790152788162, 0.033412106335163116, -0.047448933124542236, 0.03804153576493263, -0.0025254099164158106, -0.11207924783229828, 0.019641218706965446, -0.00460948096588254, -0.0742158442735672, 0.1268945336341858, 0.0407399944961071, 0.010224059224128723, -0.03741471841931343, 0.24361543357372284, -0.06653323769569397, -0.06378097087144852, -0.13251738250255585, 0.10491154342889786, -0.0027236645109951496, 0.06476365029811859, 0.023412218317389488, -0.1284150779247284, 0.005243356805294752, 0.13858191668987274, 0.12181595712900162, 0.0045748427510261536, 0.009228081442415714, 0.0518609918653965, 0.0025186820421367884, -0.06998204439878464, 0.054019294679164886, 0.06992026418447495, 0.12919506430625916, -0.07847554981708527, 0.07680778950452805, 0.0006860480643808842, -0.08370215445756912, -0.02947772853076458, 0.11312682181596756, -0.0409729965031147, 0.03491825982928276, -0.047444481402635574, 0.10916327685117722, -0.05787910893559456, -0.29412412643432617, 0.02350960113108158, -0.09588567912578583, -0.15202060341835022, -0.018367812037467957, 0.05944539234042168, -0.02624768204987049, 0.018029648810625076, 0.06971040368080139, -0.06011629104614258, 0.20098382234573364, 0.0335683599114418, -0.07864278554916382, -0.0664360448718071, 0.04837050288915634, -0.06564252078533173, 0.2949807047843933, 0.008418165147304535, 0.02863333560526371, 0.10770907253026962, -0.03253700211644173, -0.18271861970424652, 0.010723991319537163, 0.1133992001414299, -0.08056149631738663, 0.08200647681951523, 0.19000613689422607, -0.012578671798110008, 0.1209007054567337, 0.05294662341475487, -0.047376248985528946, 0.04217283055186272, -0.03389401361346245, -0.051268599927425385, -0.10752558708190918, 0.058453381061553955, -0.05909625440835953, 0.15447644889354706, 0.10152646154165268, -0.05671518296003342, -0.004550917539745569, -0.05555408447980881, 0.04875178262591362, 0.01804669201374054, 0.12263146042823792, 0.02951994352042675, -0.1865430772304535, 0.032826557755470276, -0.01144319772720337, 0.10186848044395447, -0.25588861107826233, -0.08421015739440918, 0.08833149075508118, -0.011924264021217823, -0.05105875805020332, 0.10560628771781921, 0.057650718837976456, 0.04243382066488266, -0.043439045548439026, -0.10480839014053345, -0.02186836116015911, 0.14663739502429962, -0.1469624787569046, -0.025013303384184837 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Whisper Base Bisyllabic Jyutping This model is a fine-tuned version of [openai/whisper-base](https://huggingface.co/openai/whisper-base) on the AlienKevin/cantone dataset. It achieves the following results on the evaluation set: - Loss: 0.3613 - Wer: 41.25 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 400 - training_steps: 2400 ### Training results | Training Loss | Epoch | Step | Validation Loss | Wer | |:-------------:|:-----:|:----:|:---------------:|:-------:| | 0.1093 | 0.08 | 400 | 0.3231 | 51.0417 | | 0.0389 | 0.15 | 800 | 0.2922 | 40.4861 | | 0.0237 | 0.23 | 1200 | 0.3020 | 37.7778 | | 0.0131 | 0.3 | 1600 | 0.3561 | 42.7083 | | 0.01 | 0.38 | 2000 | 0.3817 | 44.6528 | | 0.0095 | 0.46 | 2400 | 0.3613 | 41.25 | ### Framework versions - Transformers 4.38.0.dev0 - Pytorch 2.1.0 - Datasets 2.14.5 - Tokenizers 0.15.1
{"language": ["yue"], "license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["wer"], "base_model": "openai/whisper-base", "model-index": [{"name": "Whisper Base Bisyllabic Jyutping", "results": []}]}
automatic-speech-recognition
AlienKevin/whisper-base-bisyllabic-jyutping
[ "transformers", "tensorboard", "safetensors", "whisper", "automatic-speech-recognition", "generated_from_trainer", "yue", "base_model:openai/whisper-base", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2024-02-10T13:17:00+00:00
[]
[ "yue" ]
TAGS #transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #yue #base_model-openai/whisper-base #license-apache-2.0 #endpoints_compatible #region-us
Whisper Base Bisyllabic Jyutping ================================ This model is a fine-tuned version of openai/whisper-base on the AlienKevin/cantone dataset. It achieves the following results on the evaluation set: * Loss: 0.3613 * Wer: 41.25 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 1e-05 * train\_batch\_size: 32 * eval\_batch\_size: 32 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 400 * training\_steps: 2400 ### Training results ### Framework versions * Transformers 4.38.0.dev0 * Pytorch 2.1.0 * Datasets 2.14.5 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 400\n* training\\_steps: 2400", "### Training results", "### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.0\n* Datasets 2.14.5\n* Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #yue #base_model-openai/whisper-base #license-apache-2.0 #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 400\n* training\\_steps: 2400", "### Training results", "### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.0\n* Datasets 2.14.5\n* Tokenizers 0.15.1" ]
[ 71, 115, 4, 35 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #whisper #automatic-speech-recognition #generated_from_trainer #yue #base_model-openai/whisper-base #license-apache-2.0 #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 400\n* training\\_steps: 2400### Training results### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.0\n* Datasets 2.14.5\n* Tokenizers 0.15.1" ]
[ -0.10426213592290878, 0.12275201082229614, -0.0037801957223564386, 0.07633250206708908, 0.0798211321234703, -0.029333341866731644, 0.1729552298784256, 0.14572720229625702, -0.09194523841142654, 0.07240274548530579, 0.10668366402387619, 0.11803773790597916, 0.04197275638580322, 0.17875000834465027, -0.06633415073156357, -0.23476308584213257, 0.050888750702142715, 0.0065641519613564014, -0.023117095232009888, 0.10809917002916336, 0.07316985726356506, -0.12513521313667297, 0.06373448669910431, 0.019771264865994453, -0.1283138543367386, -0.010655934922397137, 0.00955644715577364, -0.09908822923898697, 0.10359487682580948, 0.004130328539758921, 0.06894415616989136, 0.060049522668123245, 0.05655074864625931, -0.20876944065093994, 0.013620133511722088, 0.05174877122044563, 0.02190019190311432, 0.06941003352403641, 0.023762507364153862, -0.01833220385015011, 0.06504448503255844, -0.07609796524047852, 0.07986147701740265, 0.03851386904716492, -0.10685697942972183, -0.3067172169685364, -0.09836445748806, 0.042181044816970825, 0.09688824415206909, 0.06950303167104721, -0.01399060059338808, 0.17402298748493195, -0.007039813324809074, 0.10398577153682709, 0.2643381953239441, -0.3258548974990845, -0.04376940056681633, -0.022772323340177536, 0.03539952263236046, 0.0693168044090271, -0.08406449109315872, 0.00391079718247056, 0.04592826962471008, 0.02845625951886177, 0.124781034886837, -0.014060144312679768, -0.013234054669737816, -0.025792790576815605, -0.13279862701892853, -0.03812725096940994, 0.13892842829227448, 0.02923795022070408, -0.06266211718320847, -0.10683263093233109, -0.06335395574569702, -0.13304059207439423, -0.063528873026371, -0.02386808581650257, 0.055719971656799316, -0.03847687691450119, -0.09679590910673141, -0.012844694778323174, -0.06565914303064346, -0.07533704489469528, -0.022839348763227463, 0.2024114429950714, 0.05186181515455246, 0.013762859627604485, -0.03877326846122742, 0.049330126494169235, -0.018677152693271637, -0.1677584946155548, -0.03425702452659607, 0.02425369992852211, -0.00034120495547540486, -0.022718828171491623, -0.03756381571292877, -0.0959797129034996, 0.05084293708205223, 0.1688309609889984, -0.09975725412368774, 0.09488622844219208, -0.03271370381116867, 0.026033341884613037, -0.09804779291152954, 0.17681053280830383, -0.013587645255029202, -0.005162627901881933, 0.017464259639382362, 0.09162065386772156, 0.0926886573433876, -0.029981939122080803, -0.08965924382209778, 0.054075874388217926, 0.0784645527601242, 0.04062902554869652, -0.04253840446472168, 0.0625525712966919, -0.0396558940410614, 0.0007024924852885306, 0.03234909847378731, -0.12502820789813995, 0.018667805939912796, 0.006348840892314911, -0.04935185983777046, -0.06960906088352203, 0.016240181401371956, 0.01577560231089592, -0.022126860916614532, 0.06753496080636978, -0.05561087653040886, 0.003330026986077428, -0.04673139378428459, -0.10788115859031677, 0.013939956203103065, -0.08642688393592834, 0.008344697766005993, -0.09910121560096741, -0.12291370332241058, -0.005212557967752218, 0.052987128496170044, -0.036062974482774734, -0.021695576608181, -0.07056474685668945, -0.10979791730642319, 0.024712946265935898, -0.014804570004343987, 0.020349862053990364, -0.07296311855316162, 0.07740014791488647, 0.05592403933405876, 0.09635181725025177, -0.004961781203746796, 0.023668058216571808, -0.08581081032752991, 0.04275776445865631, -0.2274213582277298, 0.07498938590288162, -0.09205712378025055, 0.0847901925444603, -0.10793764144182205, -0.08195272088050842, 0.03262309357523918, -0.00604186300188303, 0.0922853946685791, 0.10882949084043503, -0.1737363189458847, -0.08336862921714783, 0.22922596335411072, -0.13410194218158722, -0.14363598823547363, 0.1447015106678009, -0.0344814732670784, -0.001780231366865337, 0.07447721064090729, 0.2725692391395569, 0.054860442876815796, -0.13922089338302612, -0.005704719573259354, -0.03494682163000107, 0.03882044553756714, -0.018604135140776634, 0.06147376075387001, -0.012183165177702904, 0.06532255560159683, 0.0069874851033091545, 0.008730346336960793, 0.020073117688298225, -0.07012151181697845, -0.09103905409574509, -0.05110735073685646, -0.096641406416893, 0.012415739707648754, 0.021991843357682228, 0.053662341088056564, -0.1474495232105255, -0.10451027005910873, 0.01613522134721279, 0.09449075162410736, -0.088140107691288, 0.046734295785427094, -0.1418318748474121, 0.10215189307928085, -0.03050359897315502, -0.0042999861761927605, -0.1435442566871643, 0.04804012551903725, 0.044176600873470306, -0.05271848291158676, 0.02433173917233944, -0.07504455745220184, 0.09003335237503052, 0.06784547120332718, -0.051540639251470566, -0.03823533281683922, -0.00021234403538983315, 0.01357505563646555, -0.08492627739906311, -0.2260499894618988, -0.015391894616186619, -0.05907444655895233, 0.06480052322149277, -0.14791588485240936, 0.0468423068523407, 0.06784062832593918, 0.11162805557250977, 0.06035838648676872, -0.030082425102591515, 0.013884693384170532, 0.0519946925342083, -0.01441364735364914, -0.06794777512550354, 0.0367654412984848, 0.03335561975836754, -0.10217148065567017, 0.0260304044932127, -0.18814386427402496, 0.14653748273849487, 0.1401168406009674, 0.014045337215065956, -0.052597276866436005, 0.016880810260772705, -0.040860142558813095, -0.02507067658007145, -0.008842934854328632, 0.02918045036494732, 0.19171515107154846, -0.0024545451160520315, 0.14138789474964142, -0.10786889493465424, -0.04661476984620094, 0.043814267963171005, -0.04931773245334625, -0.007974378764629364, 0.10064683854579926, -0.028695236891508102, -0.0774577185511589, 0.11375822126865387, 0.11366458237171173, -0.07147242873907089, 0.13155414164066315, -0.08285629004240036, -0.04372049495577812, -0.0258018895983696, 0.02594623900949955, 0.0362650528550148, 0.11029598861932755, -0.11488743126392365, -0.02843657322227955, 0.01213778369128704, 0.005896815564483404, 0.0047744270414114, -0.1943926215171814, 0.00460265576839447, 0.042083945125341415, -0.07752026617527008, -0.027569493278861046, -0.0025769711937755346, -0.010313596576452255, 0.0928255245089531, 0.003002353711053729, -0.07666587084531784, 0.016505323350429535, -0.014647829346358776, -0.06392642110586166, 0.17580534517765045, -0.10377925634384155, -0.14710545539855957, -0.09352477639913559, -0.07079649716615677, -0.034119799733161926, 0.018053123727440834, 0.08628649264574051, -0.11434534192085266, -0.04120476916432381, -0.11467505991458893, -0.016701407730579376, 0.031528353691101074, 0.04137613624334335, 0.09555580466985703, 0.00362488254904747, 0.09170212596654892, -0.1109619066119194, -0.017875835299491882, -0.04837265983223915, 0.0010482233483344316, 0.02547193504869938, 0.03492721915245056, 0.08545809239149094, 0.13556134700775146, -0.014011994004249573, 0.034621018916368484, -0.038121677935123444, 0.20397628843784332, -0.08155226707458496, -0.027635565027594566, 0.14109812676906586, -0.02256818301975727, 0.06927650421857834, 0.1497085839509964, 0.04526245594024658, -0.11722838878631592, -0.004149266984313726, 0.0008581639849580824, -0.05027065798640251, -0.21427203714847565, -0.0264755766838789, -0.029898688197135925, 0.02421390637755394, 0.07206068933010101, 0.038931164890527725, 0.04308221489191055, 0.028364019468426704, 0.01592552661895752, -0.0018713001627475023, 0.024243580177426338, 0.09772657603025436, 0.10130396485328674, 0.04357481747865677, 0.12634916603565216, -0.05920005589723587, -0.03495559096336365, 0.027037741616368294, 0.0036240187473595142, 0.20587074756622314, 0.01244957372546196, 0.17621541023254395, 0.05289888009428978, 0.11891621351242065, 0.028363678604364395, 0.030852805823087692, -0.019037943333387375, -0.02702152170240879, 0.0035137678496539593, -0.06220708414912224, -0.036213260143995285, 0.042125996202230453, -0.05484291911125183, 0.04848071560263634, -0.08822286128997803, 0.06241068243980408, 0.07233153283596039, 0.31107598543167114, 0.05463090166449547, -0.31751012802124023, -0.10985005646944046, 0.017408868297934532, -0.07093960791826248, -0.0004979399964213371, 0.05020364746451378, 0.13798299431800842, -0.03070143423974514, 0.08935469388961792, -0.05935129150748253, 0.08074509352445602, -0.06118073686957359, 0.040337007492780685, 0.024000177159905434, 0.09697495400905609, -0.013115228153765202, 0.012764722108840942, -0.25845012068748474, 0.29094216227531433, 0.01857699826359749, 0.10690752416849136, -0.04562239348888397, 0.019210146740078926, 0.023310858756303787, 0.03748686611652374, 0.099390409886837, -0.02048138715326786, -0.1438615471124649, -0.13525916635990143, -0.09992112219333649, 0.021723918616771698, 0.12237044423818588, 0.03254605084657669, 0.09665864706039429, -0.009676489047706127, -0.01597316563129425, 0.048281989991664886, -0.08045767992734909, -0.05727630481123924, -0.07154110819101334, 0.0006490356754511595, 0.07891786843538284, -0.028879310935735703, -0.08579261600971222, -0.08602281659841537, -0.0859694704413414, 0.14528687298297882, -0.03689483180642128, -0.05804058536887169, -0.10290540754795074, -0.014499198645353317, 0.0778370052576065, -0.07086396217346191, 0.0444992333650589, 0.009927595965564251, 0.09836194664239883, 0.00035393217694945633, -0.05808800086379051, 0.11720174551010132, -0.06928784400224686, -0.18888133764266968, -0.041709475219249725, 0.15787452459335327, 0.018339386209845543, 0.054036129266023636, 0.002950244816020131, 0.03424771875143051, 0.005348611623048782, -0.06755676865577698, 0.06422771513462067, 0.004926105495542288, 0.04117027297616005, -0.030049879103899002, 0.009013253264129162, -0.022614656016230583, -0.0918918028473854, -0.017665190622210503, 0.14204181730747223, 0.30448615550994873, -0.0833963081240654, 0.10007946193218231, 0.08976061642169952, -0.03971267491579056, -0.17796102166175842, -0.008386112749576569, 0.04127410054206848, -0.007769498974084854, -0.02446652017533779, -0.16870030760765076, 0.05110098794102669, 0.05235455185174942, -0.04414847493171692, 0.06836146861314774, -0.2862827479839325, -0.1442178338766098, 0.11747579276561737, 0.12135890871286392, 0.1098436564207077, -0.1521163284778595, -0.061209291219711304, -0.02283005230128765, -0.08113076537847519, 0.049502547830343246, -0.13663053512573242, 0.11355524510145187, 0.00006314666097750887, 0.04822588339447975, 0.017526380717754364, -0.058425042778253555, 0.11728058010339737, -0.01668662764132023, 0.10010537505149841, -0.06477750837802887, 0.02882647141814232, 0.06144241616129875, -0.07634776085615158, 0.050560567528009415, -0.12866903841495514, 0.052747100591659546, -0.026712996885180473, -0.02465640939772129, -0.05351182818412781, 0.004176491405814886, -0.006162256468087435, -0.04046022891998291, -0.03856927156448364, 0.009585953317582607, 0.07295756042003632, -0.010718273930251598, 0.1878325492143631, -0.011634895578026772, 0.15036527812480927, 0.16876909136772156, 0.11894220858812332, -0.09975907951593399, -0.011871017515659332, 0.021259533241391182, -0.04091377928853035, 0.065274678170681, -0.17606358230113983, 0.05686235800385475, 0.1051478385925293, 0.0217781700193882, 0.11180844902992249, 0.05259455367922783, -0.06916418671607971, 0.037428565323352814, 0.05835644528269768, -0.14307866990566254, -0.16140080988407135, 0.021620701998472214, 0.03543030470609665, -0.09129703044891357, 0.1079542264342308, 0.1265895813703537, -0.06466598063707352, 0.00001918823181767948, -0.007983296178281307, 0.030152371153235435, -0.03287846967577934, 0.1901274472475052, 0.032713793218135834, 0.0636029914021492, -0.11323633790016174, 0.1067911833524704, 0.01774457097053528, -0.08885170519351959, 0.07994105666875839, 0.08130109310150146, -0.09812904894351959, -0.0217281561344862, 0.025967368856072426, 0.12409316748380661, 0.050775397568941116, -0.06295686215162277, -0.16330967843532562, -0.13137033581733704, 0.07257235795259476, 0.22109180688858032, 0.06055011972784996, 0.009959620423614979, -0.027353331446647644, 0.026980405673384666, -0.11275032162666321, 0.10124469548463821, 0.03861482813954353, 0.057553548365831375, -0.14580748975276947, 0.12495463341474533, 0.004824153613299131, 0.017801744863390923, -0.027510300278663635, 0.011623496189713478, -0.11718348413705826, 0.023153390735387802, -0.15203812718391418, 0.02403460443019867, -0.056393884122371674, 0.0052522579208016396, -0.0037441577296704054, -0.052687738090753555, -0.07192499935626984, 0.04646596685051918, -0.11110790073871613, -0.015435914508998394, 0.021647101268172264, 0.021585097536444664, -0.12038231641054153, -0.027483858168125153, 0.01188687328249216, -0.09271314740180969, 0.08749043941497803, 0.05896971374750137, -0.03394899144768715, 0.04108452424407005, -0.09272348880767822, -0.007991083897650242, 0.0818299651145935, -0.0023510681930929422, 0.03502264246344566, -0.12568002939224243, -0.038838814944028854, 0.007815510034561157, 0.01457591075450182, 0.014099881052970886, 0.1279752403497696, -0.0986655056476593, 0.023515932261943817, -0.014116339385509491, -0.048047199845314026, -0.05359445512294769, 0.031552504748106, 0.12096333503723145, 0.007162903901189566, 0.15983423590660095, -0.10458271205425262, 0.01416968647390604, -0.18431414663791656, -0.000718550814781338, 0.004984940402209759, -0.13492676615715027, -0.10202138870954514, -0.01044860202819109, 0.07377278059720993, -0.08079858124256134, 0.1333044320344925, -0.03880058974027634, -0.006620100233703852, 0.05006822943687439, -0.07171615958213806, -0.019381407648324966, 0.044215328991413116, 0.19892680644989014, 0.024921681731939316, -0.04620997980237007, 0.06538476049900055, -0.009963546879589558, 0.0903061106801033, 0.0723591074347496, 0.1436651051044464, 0.18051102757453918, 0.06289727985858917, 0.10934232175350189, 0.06921737641096115, -0.028415514156222343, -0.18581731617450714, 0.0709533765912056, -0.06587576121091843, 0.11393749713897705, -0.005550963804125786, 0.177882120013237, 0.15532565116882324, -0.12007274478673935, 0.056022267788648605, -0.026385486125946045, -0.07286999374628067, -0.12934072315692902, -0.08589047938585281, -0.0987272635102272, -0.16655077040195465, 0.012471574358642101, -0.10637348145246506, 0.04026421532034874, 0.04525375738739967, 0.024328580126166344, 0.016944577917456627, 0.1538149118423462, 0.0018972904654219747, 0.04167448356747627, 0.07872927933931351, -0.0142739312723279, -0.07396847009658813, -0.013703000731766224, -0.09465961158275604, 0.05375862121582031, -0.0006817029207013547, 0.03495728224515915, -0.006414658389985561, -0.04798216000199318, 0.051873188465833664, -0.02998223714530468, -0.11439965665340424, 0.026463687419891357, 0.022060146555304527, 0.06606777757406235, 0.020221898332238197, 0.038084566593170166, -0.02467699721455574, 0.003116678912192583, 0.21630972623825073, -0.09538877010345459, -0.09044148027896881, -0.1276949942111969, 0.24140521883964539, 0.011404714547097683, -0.0025625312700867653, 0.005047300364822149, -0.07770999521017075, -0.026915600523352623, 0.17895598709583282, 0.19553475081920624, -0.012596541084349155, 0.01152481883764267, -0.0435384139418602, -0.004829478915780783, -0.062394194304943085, 0.06838254630565643, 0.11945269256830215, 0.01845015026628971, -0.03983727842569351, -0.047526001930236816, -0.03741832077503204, -0.035235658288002014, -0.05650952458381653, 0.06946488469839096, -0.0021916376426815987, -0.014627992175519466, -0.04720575362443924, 0.057799406349658966, -0.06657852232456207, -0.10492191463708878, 0.026157362386584282, -0.20860116183757782, -0.14518509805202484, 0.011648857034742832, 0.07490617781877518, 0.029544321820139885, 0.04207531735301018, 0.018626488745212555, -0.0024159825406968594, 0.06366413086652756, -0.011796976439654827, -0.061497513204813004, -0.07113872468471527, 0.07396253198385239, -0.1789846420288086, 0.20300643146038055, -0.024156538769602776, 0.03518176078796387, 0.12231888622045517, 0.05068983510136604, -0.08223812282085419, 0.09128695726394653, 0.04655393958091736, -0.12099270522594452, -0.0018987111980095506, 0.15299706161022186, -0.044827692210674286, 0.12346243113279343, 0.05683360621333122, -0.09919073432683945, -0.02091699279844761, -0.06514785438776016, -0.0640207976102829, -0.04818464815616608, -0.029333699494600296, -0.057126302272081375, 0.1119452640414238, 0.16582635045051575, -0.058214373886585236, 0.011383198201656342, -0.029133982956409454, 0.021789245307445526, 0.07028224319219589, -0.028775731101632118, -0.0403081513941288, -0.28323283791542053, 0.015320315957069397, 0.0570388063788414, 0.001830228604376316, -0.2504112124443054, -0.08910390734672546, -0.001860021729953587, -0.027149846777319908, -0.07916188985109329, 0.06906954199075699, 0.12761469185352325, 0.04970403015613556, -0.06324838101863861, -0.06639706343412399, -0.024677244946360588, 0.17359164357185364, -0.14732621610164642, -0.07303345203399658 ]
null
null
transformers
# LinkBERT: Fine-tuned BERT for Natural Link Prediction LinkBERT is an advanced fine-tuned version of the [bert-large-cased](https://huggingface.co/google-bert/bert-large-cased) model developed by [Dejan Marketing](https://dejanmarketing.com/). The model is designed to predict natural link placement within web content. This binary classification model excels in identifying distinct token ranges that web authors are likely to choose as anchor text for links. By analyzing never-before-seen texts, LinkBERT can predict areas within the content where links might naturally occur, effectively simulating web author behavior in link creation. # Online Demo Online demo of this model is available at https://linkbert.com/ ## Applications of LinkBERT LinkBERT's applications are vast and diverse, tailored to enhance both the efficiency and quality of web content creation and analysis: - **Anchor Text Suggestion:** Acts as a mechanism during internal link optimization, suggesting potential anchor texts to web authors. - **Evaluation of Existing Links:** Assesses the naturalness of link placements within existing content, aiding in the refinement of web pages. - **Link Placement Guide:** Offers guidance to link builders by suggesting optimal placement for links within content. - **Anchor Text Idea Generator:** Provides creative anchor text suggestions to enrich content and improve SEO strategies. - **Spam and Inorganic SEO Detection:** Helps identify unnatural link patterns, contributing to the detection of spam and inorganic SEO tactics. ## Training and Performance LinkBERT was fine-tuned on a dataset of organic web content and editorial links. The training involved preprocessing web content, annotating links with temporary markup for clear distinction, and employing a specialized tokenization process to prepare the data for model training. In adition to commonly available data sources such as Wikipedia, additional training data was also sourced from: - [Owayo USA](https://www.owayo.com/) - [Owayo Germany](https://www.owayo.de/) - [Compare the Market](https://www.comparethemarket.com.au/) - [Leonardo AI](https://leonardo.ai/) - [The Wests Group](https://mywests.com.au/) - [Zendesk](https://www.zendesk.com/) - [Australia Post](https://auspost.com.au/) - [Probe Group](https://www.probecx.com/) - [Monarch Institute](https://www.monarch.edu.au/) - And many others... ### Training Highlights: - **Dataset:** Custom organic web content with editorial links. - **Preprocessing:** Links annotated with `[START_LINK]` and `[END_LINK]` markup. - **Tokenization:** Utilized input_ids, token_type_ids, attention_mask, and labels for model training, with a unique labeling system to differentiate between link/anchor text and plain text. ### Technical Specifications: - **Batch Size:** 10, with class weights adjusted to address class imbalance between link and plain text. - **Optimizer:** AdamW with a learning rate of 5e-5. - **Epochs:** 5, incorporating gradient accumulation and warmup steps to optimize training outcomes. - **Hardware:** 1 x RTX4090 24GB VRAM - **Duration:** 32 hours ## Utilization and Integration LinkBERT is positioned as a powerful tool for content creators, SEO specialists, and webmasters, offering unparalleled support in optimizing web content for both user engagement and search engine recognition. Its predictive capabilities not only streamline the content creation process but also offer insights into the natural integration of links, enhancing the overall quality and relevance of web content. ## Accessibility LinkBERT leverages the robust architecture of bert-large-cased, enhancing it with capabilities specifically tailored for web content analysis. This model represents a significant advancement in the understanding and generation of web content, providing a nuanced approach to natural link prediction and anchor text suggestion. --- # BERT large model (cased) Pretrained model on English language using a masked language modeling (MLM) objective. It was introduced in [this paper](https://arxiv.org/abs/1810.04805) and first released in [this repository](https://github.com/google-research/bert). This model is cased: it makes a difference between english and English. Disclaimer: The team releasing BERT did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description BERT is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it was pretrained with two objectives: - Masked language modeling (MLM): taking a sentence, the model randomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict the masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one after the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to learn a bidirectional representation of the sentence. - Next sentence prediction (NSP): the models concatenates two masked sentences as inputs during pretraining. Sometimes they correspond to sentences that were next to each other in the original text, sometimes not. The model then has to predict if the two sentences were following each other or not. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. This model has the following configuration: - 24-layer - 1024 hidden dimension - 16 attention heads - 336M parameters.
{"language": "en", "license": "bigscience-bloom-rail-1.0", "datasets": ["bookcorpus", "wikipedia"], "pipeline_tag": "token-classification", "widget": [{"text": "LinkBERT is an advanced fine-tuned version of the bert-large-cased model developed by Dejan Marketing. The model is designed to predict natural link placement within web content."}]}
token-classification
dejanseo/LinkBERT
[ "transformers", "pytorch", "bert", "fill-mask", "token-classification", "en", "dataset:bookcorpus", "dataset:wikipedia", "arxiv:1810.04805", "license:bigscience-bloom-rail-1.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-10T13:18:00+00:00
[ "1810.04805" ]
[ "en" ]
TAGS #transformers #pytorch #bert #fill-mask #token-classification #en #dataset-bookcorpus #dataset-wikipedia #arxiv-1810.04805 #license-bigscience-bloom-rail-1.0 #autotrain_compatible #endpoints_compatible #region-us
# LinkBERT: Fine-tuned BERT for Natural Link Prediction LinkBERT is an advanced fine-tuned version of the bert-large-cased model developed by Dejan Marketing. The model is designed to predict natural link placement within web content. This binary classification model excels in identifying distinct token ranges that web authors are likely to choose as anchor text for links. By analyzing never-before-seen texts, LinkBERT can predict areas within the content where links might naturally occur, effectively simulating web author behavior in link creation. # Online Demo Online demo of this model is available at URL ## Applications of LinkBERT LinkBERT's applications are vast and diverse, tailored to enhance both the efficiency and quality of web content creation and analysis: - Anchor Text Suggestion: Acts as a mechanism during internal link optimization, suggesting potential anchor texts to web authors. - Evaluation of Existing Links: Assesses the naturalness of link placements within existing content, aiding in the refinement of web pages. - Link Placement Guide: Offers guidance to link builders by suggesting optimal placement for links within content. - Anchor Text Idea Generator: Provides creative anchor text suggestions to enrich content and improve SEO strategies. - Spam and Inorganic SEO Detection: Helps identify unnatural link patterns, contributing to the detection of spam and inorganic SEO tactics. ## Training and Performance LinkBERT was fine-tuned on a dataset of organic web content and editorial links. The training involved preprocessing web content, annotating links with temporary markup for clear distinction, and employing a specialized tokenization process to prepare the data for model training. In adition to commonly available data sources such as Wikipedia, additional training data was also sourced from: - Owayo USA - Owayo Germany - Compare the Market - Leonardo AI - The Wests Group - Zendesk - Australia Post - Probe Group - Monarch Institute - And many others... ### Training Highlights: - Dataset: Custom organic web content with editorial links. - Preprocessing: Links annotated with '[START_LINK]' and '[END_LINK]' markup. - Tokenization: Utilized input_ids, token_type_ids, attention_mask, and labels for model training, with a unique labeling system to differentiate between link/anchor text and plain text. ### Technical Specifications: - Batch Size: 10, with class weights adjusted to address class imbalance between link and plain text. - Optimizer: AdamW with a learning rate of 5e-5. - Epochs: 5, incorporating gradient accumulation and warmup steps to optimize training outcomes. - Hardware: 1 x RTX4090 24GB VRAM - Duration: 32 hours ## Utilization and Integration LinkBERT is positioned as a powerful tool for content creators, SEO specialists, and webmasters, offering unparalleled support in optimizing web content for both user engagement and search engine recognition. Its predictive capabilities not only streamline the content creation process but also offer insights into the natural integration of links, enhancing the overall quality and relevance of web content. ## Accessibility LinkBERT leverages the robust architecture of bert-large-cased, enhancing it with capabilities specifically tailored for web content analysis. This model represents a significant advancement in the understanding and generation of web content, providing a nuanced approach to natural link prediction and anchor text suggestion. --- # BERT large model (cased) Pretrained model on English language using a masked language modeling (MLM) objective. It was introduced in this paper and first released in this repository. This model is cased: it makes a difference between english and English. Disclaimer: The team releasing BERT did not write a model card for this model so this model card has been written by the Hugging Face team. ## Model description BERT is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it was pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it was pretrained with two objectives: - Masked language modeling (MLM): taking a sentence, the model randomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict the masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the words one after the other, or from autoregressive models like GPT which internally mask the future tokens. It allows the model to learn a bidirectional representation of the sentence. - Next sentence prediction (NSP): the models concatenates two masked sentences as inputs during pretraining. Sometimes they correspond to sentences that were next to each other in the original text, sometimes not. The model then has to predict if the two sentences were following each other or not. This way, the model learns an inner representation of the English language that can then be used to extract features useful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard classifier using the features produced by the BERT model as inputs. This model has the following configuration: - 24-layer - 1024 hidden dimension - 16 attention heads - 336M parameters.
[ "# LinkBERT: Fine-tuned BERT for Natural Link Prediction\n\nLinkBERT is an advanced fine-tuned version of the bert-large-cased model developed by Dejan Marketing. The model is designed to predict natural link placement within web content. This binary classification model excels in identifying distinct token ranges that web authors are likely to choose as anchor text for links. By analyzing never-before-seen texts, LinkBERT can predict areas within the content where links might naturally occur, effectively simulating web author behavior in link creation.", "# Online Demo\n\nOnline demo of this model is available at URL", "## Applications of LinkBERT\n\nLinkBERT's applications are vast and diverse, tailored to enhance both the efficiency and quality of web content creation and analysis:\n\n- Anchor Text Suggestion: Acts as a mechanism during internal link optimization, suggesting potential anchor texts to web authors.\n- Evaluation of Existing Links: Assesses the naturalness of link placements within existing content, aiding in the refinement of web pages.\n- Link Placement Guide: Offers guidance to link builders by suggesting optimal placement for links within content.\n- Anchor Text Idea Generator: Provides creative anchor text suggestions to enrich content and improve SEO strategies.\n- Spam and Inorganic SEO Detection: Helps identify unnatural link patterns, contributing to the detection of spam and inorganic SEO tactics.", "## Training and Performance\n\nLinkBERT was fine-tuned on a dataset of organic web content and editorial links. The training involved preprocessing web content, annotating links with temporary markup for clear distinction, and employing a specialized tokenization process to prepare the data for model training. In adition to commonly available data sources such as Wikipedia, additional training data was also sourced from:\n\n- Owayo USA\n- Owayo Germany\n- Compare the Market\n- Leonardo AI\n- The Wests Group\n- Zendesk\n- Australia Post\n- Probe Group\n- Monarch Institute\n- And many others...", "### Training Highlights:\n\n- Dataset: Custom organic web content with editorial links.\n- Preprocessing: Links annotated with '[START_LINK]' and '[END_LINK]' markup.\n- Tokenization: Utilized input_ids, token_type_ids, attention_mask, and labels for model training, with a unique labeling system to differentiate between link/anchor text and plain text.", "### Technical Specifications:\n\n- Batch Size: 10, with class weights adjusted to address class imbalance between link and plain text.\n- Optimizer: AdamW with a learning rate of 5e-5.\n- Epochs: 5, incorporating gradient accumulation and warmup steps to optimize training outcomes.\n- Hardware: 1 x RTX4090 24GB VRAM\n- Duration: 32 hours", "## Utilization and Integration\n\nLinkBERT is positioned as a powerful tool for content creators, SEO specialists, and webmasters, offering unparalleled support in optimizing web content for both user engagement and search engine recognition. Its predictive capabilities not only streamline the content creation process but also offer insights into the natural integration of links, enhancing the overall quality and relevance of web content.", "## Accessibility\n\nLinkBERT leverages the robust architecture of bert-large-cased, enhancing it with capabilities specifically tailored for web content analysis. This model represents a significant advancement in the understanding and generation of web content, providing a nuanced approach to natural link prediction and anchor text suggestion.\n\n---", "# BERT large model (cased)\n\nPretrained model on English language using a masked language modeling (MLM) objective. It was introduced in\nthis paper and first released in\nthis repository. This model is cased: it makes a difference\nbetween english and English.\n\nDisclaimer: The team releasing BERT did not write a model card for this model so this model card has been written by\nthe Hugging Face team.", "## Model description\n\nBERT is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it\nwas pretrained with two objectives:\n\n- Masked language modeling (MLM): taking a sentence, the model randomly masks 15% of the words in the input then run\n the entire masked sentence through the model and has to predict the masked words. This is different from traditional\n recurrent neural networks (RNNs) that usually see the words one after the other, or from autoregressive models like\n GPT which internally mask the future tokens. It allows the model to learn a bidirectional representation of the\n sentence.\n- Next sentence prediction (NSP): the models concatenates two masked sentences as inputs during pretraining. Sometimes\n they correspond to sentences that were next to each other in the original text, sometimes not. The model then has to\n predict if the two sentences were following each other or not.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.\n\nThis model has the following configuration:\n\n- 24-layer\n- 1024 hidden dimension\n- 16 attention heads\n- 336M parameters." ]
[ "TAGS\n#transformers #pytorch #bert #fill-mask #token-classification #en #dataset-bookcorpus #dataset-wikipedia #arxiv-1810.04805 #license-bigscience-bloom-rail-1.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# LinkBERT: Fine-tuned BERT for Natural Link Prediction\n\nLinkBERT is an advanced fine-tuned version of the bert-large-cased model developed by Dejan Marketing. The model is designed to predict natural link placement within web content. This binary classification model excels in identifying distinct token ranges that web authors are likely to choose as anchor text for links. By analyzing never-before-seen texts, LinkBERT can predict areas within the content where links might naturally occur, effectively simulating web author behavior in link creation.", "# Online Demo\n\nOnline demo of this model is available at URL", "## Applications of LinkBERT\n\nLinkBERT's applications are vast and diverse, tailored to enhance both the efficiency and quality of web content creation and analysis:\n\n- Anchor Text Suggestion: Acts as a mechanism during internal link optimization, suggesting potential anchor texts to web authors.\n- Evaluation of Existing Links: Assesses the naturalness of link placements within existing content, aiding in the refinement of web pages.\n- Link Placement Guide: Offers guidance to link builders by suggesting optimal placement for links within content.\n- Anchor Text Idea Generator: Provides creative anchor text suggestions to enrich content and improve SEO strategies.\n- Spam and Inorganic SEO Detection: Helps identify unnatural link patterns, contributing to the detection of spam and inorganic SEO tactics.", "## Training and Performance\n\nLinkBERT was fine-tuned on a dataset of organic web content and editorial links. The training involved preprocessing web content, annotating links with temporary markup for clear distinction, and employing a specialized tokenization process to prepare the data for model training. In adition to commonly available data sources such as Wikipedia, additional training data was also sourced from:\n\n- Owayo USA\n- Owayo Germany\n- Compare the Market\n- Leonardo AI\n- The Wests Group\n- Zendesk\n- Australia Post\n- Probe Group\n- Monarch Institute\n- And many others...", "### Training Highlights:\n\n- Dataset: Custom organic web content with editorial links.\n- Preprocessing: Links annotated with '[START_LINK]' and '[END_LINK]' markup.\n- Tokenization: Utilized input_ids, token_type_ids, attention_mask, and labels for model training, with a unique labeling system to differentiate between link/anchor text and plain text.", "### Technical Specifications:\n\n- Batch Size: 10, with class weights adjusted to address class imbalance between link and plain text.\n- Optimizer: AdamW with a learning rate of 5e-5.\n- Epochs: 5, incorporating gradient accumulation and warmup steps to optimize training outcomes.\n- Hardware: 1 x RTX4090 24GB VRAM\n- Duration: 32 hours", "## Utilization and Integration\n\nLinkBERT is positioned as a powerful tool for content creators, SEO specialists, and webmasters, offering unparalleled support in optimizing web content for both user engagement and search engine recognition. Its predictive capabilities not only streamline the content creation process but also offer insights into the natural integration of links, enhancing the overall quality and relevance of web content.", "## Accessibility\n\nLinkBERT leverages the robust architecture of bert-large-cased, enhancing it with capabilities specifically tailored for web content analysis. This model represents a significant advancement in the understanding and generation of web content, providing a nuanced approach to natural link prediction and anchor text suggestion.\n\n---", "# BERT large model (cased)\n\nPretrained model on English language using a masked language modeling (MLM) objective. It was introduced in\nthis paper and first released in\nthis repository. This model is cased: it makes a difference\nbetween english and English.\n\nDisclaimer: The team releasing BERT did not write a model card for this model so this model card has been written by\nthe Hugging Face team.", "## Model description\n\nBERT is a transformers model pretrained on a large corpus of English data in a self-supervised fashion. This means it\nwas pretrained on the raw texts only, with no humans labelling them in any way (which is why it can use lots of\npublicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it\nwas pretrained with two objectives:\n\n- Masked language modeling (MLM): taking a sentence, the model randomly masks 15% of the words in the input then run\n the entire masked sentence through the model and has to predict the masked words. This is different from traditional\n recurrent neural networks (RNNs) that usually see the words one after the other, or from autoregressive models like\n GPT which internally mask the future tokens. It allows the model to learn a bidirectional representation of the\n sentence.\n- Next sentence prediction (NSP): the models concatenates two masked sentences as inputs during pretraining. Sometimes\n they correspond to sentences that were next to each other in the original text, sometimes not. The model then has to\n predict if the two sentences were following each other or not.\n\nThis way, the model learns an inner representation of the English language that can then be used to extract features\nuseful for downstream tasks: if you have a dataset of labeled sentences for instance, you can train a standard\nclassifier using the features produced by the BERT model as inputs.\n\nThis model has the following configuration:\n\n- 24-layer\n- 1024 hidden dimension\n- 16 attention heads\n- 336M parameters." ]
[ 78, 129, 12, 183, 129, 96, 88, 87, 74, 92, 359 ]
[ "passage: TAGS\n#transformers #pytorch #bert #fill-mask #token-classification #en #dataset-bookcorpus #dataset-wikipedia #arxiv-1810.04805 #license-bigscience-bloom-rail-1.0 #autotrain_compatible #endpoints_compatible #region-us \n# LinkBERT: Fine-tuned BERT for Natural Link Prediction\n\nLinkBERT is an advanced fine-tuned version of the bert-large-cased model developed by Dejan Marketing. The model is designed to predict natural link placement within web content. This binary classification model excels in identifying distinct token ranges that web authors are likely to choose as anchor text for links. By analyzing never-before-seen texts, LinkBERT can predict areas within the content where links might naturally occur, effectively simulating web author behavior in link creation.# Online Demo\n\nOnline demo of this model is available at URL## Applications of LinkBERT\n\nLinkBERT's applications are vast and diverse, tailored to enhance both the efficiency and quality of web content creation and analysis:\n\n- Anchor Text Suggestion: Acts as a mechanism during internal link optimization, suggesting potential anchor texts to web authors.\n- Evaluation of Existing Links: Assesses the naturalness of link placements within existing content, aiding in the refinement of web pages.\n- Link Placement Guide: Offers guidance to link builders by suggesting optimal placement for links within content.\n- Anchor Text Idea Generator: Provides creative anchor text suggestions to enrich content and improve SEO strategies.\n- Spam and Inorganic SEO Detection: Helps identify unnatural link patterns, contributing to the detection of spam and inorganic SEO tactics.", "passage: ## Training and Performance\n\nLinkBERT was fine-tuned on a dataset of organic web content and editorial links. The training involved preprocessing web content, annotating links with temporary markup for clear distinction, and employing a specialized tokenization process to prepare the data for model training. In adition to commonly available data sources such as Wikipedia, additional training data was also sourced from:\n\n- Owayo USA\n- Owayo Germany\n- Compare the Market\n- Leonardo AI\n- The Wests Group\n- Zendesk\n- Australia Post\n- Probe Group\n- Monarch Institute\n- And many others...### Training Highlights:\n\n- Dataset: Custom organic web content with editorial links.\n- Preprocessing: Links annotated with '[START_LINK]' and '[END_LINK]' markup.\n- Tokenization: Utilized input_ids, token_type_ids, attention_mask, and labels for model training, with a unique labeling system to differentiate between link/anchor text and plain text.### Technical Specifications:\n\n- Batch Size: 10, with class weights adjusted to address class imbalance between link and plain text.\n- Optimizer: AdamW with a learning rate of 5e-5.\n- Epochs: 5, incorporating gradient accumulation and warmup steps to optimize training outcomes.\n- Hardware: 1 x RTX4090 24GB VRAM\n- Duration: 32 hours## Utilization and Integration\n\nLinkBERT is positioned as a powerful tool for content creators, SEO specialists, and webmasters, offering unparalleled support in optimizing web content for both user engagement and search engine recognition. Its predictive capabilities not only streamline the content creation process but also offer insights into the natural integration of links, enhancing the overall quality and relevance of web content.## Accessibility\n\nLinkBERT leverages the robust architecture of bert-large-cased, enhancing it with capabilities specifically tailored for web content analysis. This model represents a significant advancement in the understanding and generation of web content, providing a nuanced approach to natural link prediction and anchor text suggestion.\n\n---# BERT large model (cased)\n\nPretrained model on English language using a masked language modeling (MLM) objective. It was introduced in\nthis paper and first released in\nthis repository. This model is cased: it makes a difference\nbetween english and English.\n\nDisclaimer: The team releasing BERT did not write a model card for this model so this model card has been written by\nthe Hugging Face team." ]
[ -0.06433500349521637, 0.09201404452323914, -0.0019514831947162747, 0.018913736566901207, 0.06191550940275192, -0.03604750707745552, 0.04942135512828827, 0.040816426277160645, 0.0159296914935112, 0.05344151705503464, -0.07544127106666565, -0.06016024947166443, 0.07674199342727661, 0.06464492529630661, 0.052351269870996475, -0.21509864926338196, 0.04105275869369507, -0.016897859051823616, -0.008523978292942047, 0.07715651392936707, 0.0934496521949768, -0.08208920806646347, 0.07215403765439987, 0.07656539976596832, 0.043317340314388275, 0.0034991586580872536, -0.05912300944328308, -0.06534554809331894, 0.07477951049804688, 0.028254413977265358, 0.07458983361721039, -0.009608075022697449, 0.046313636004924774, -0.19420969486236572, 0.012918040156364441, 0.06379221379756927, 0.027647461742162704, 0.07062701880931854, 0.08215964585542679, -0.0017895139753818512, 0.04675842076539993, 0.021070312708616257, 0.09376204013824463, 0.05070751905441284, -0.07191358506679535, -0.020473161712288857, -0.04397594556212425, 0.07108314335346222, 0.041578590869903564, 0.09177614748477936, -0.021776219829916954, 0.10465402901172638, -0.0749814435839653, 0.004278987646102905, 0.05404474586248398, -0.1292053461074829, -0.008599304594099522, -0.034048326313495636, 0.0372559130191803, 0.04872077703475952, -0.07652384042739868, -0.0038708215579390526, 0.02778070792555809, 0.0034269585739821196, 0.1084061861038208, 0.019366735592484474, 0.042527347803115845, -0.03455951809883118, -0.10344457626342773, -0.001931533683091402, 0.15059460699558258, 0.08111576735973358, -0.08237378299236298, -0.16696825623512268, -0.02167566306889057, 0.12725409865379333, -0.020374445244669914, -0.024743402376770973, 0.037453245371580124, 0.0367305651307106, 0.060364123433828354, -0.11297982931137085, -0.08370532095432281, 0.01968945749104023, 0.0009501632302999496, 0.16092516481876373, 0.019221041351556778, 0.010808628052473068, 0.0066202376037836075, 0.09775571525096893, 0.012403847649693489, -0.10206697881221771, 0.015287095680832863, -0.05642446130514145, -0.012716625817120075, -0.01875673420727253, -0.043474502861499786, -0.14486154913902283, -0.03302167356014252, 0.05141767859458923, 0.01853703334927559, 0.02407829649746418, -0.03722638264298439, -0.0031532752327620983, 0.0386706180870533, 0.1660459190607071, -0.024994276463985443, -0.09715114533901215, 0.012226229533553123, 0.10007722675800323, 0.010538868606090546, -0.023908201605081558, -0.02926388755440712, 0.03671533241868019, 0.01641123555600643, -0.01814064383506775, 0.012868167832493782, 0.04243788495659828, -0.07719061523675919, -0.04458773136138916, 0.1812865287065506, -0.09148889780044556, 0.033017974346876144, 0.026237860321998596, -0.056137703359127045, 0.012744845822453499, 0.04530151188373566, 0.04268236458301544, -0.10482002049684525, 0.04380282759666443, -0.06976892799139023, -0.0217897966504097, -0.08072225004434586, -0.09196056425571442, 0.026722049340605736, -0.01679682731628418, -0.0383506715297699, -0.06270664930343628, -0.1332325041294098, -0.048235710710287094, 0.03161998093128204, -0.015248949639499187, -0.03197642043232918, -0.01940135285258293, -0.0003942332696169615, -0.0011382568627595901, 0.02704283595085144, -0.004296503961086273, -0.03630648925900459, -0.017389174550771713, -0.08779701590538025, -0.0092283571138978, 0.02903093956410885, 0.025512689724564552, -0.1282716691493988, -0.01351885125041008, -0.1279180943965912, 0.12478455901145935, -0.07611793279647827, 0.033501509577035904, -0.06474220752716064, -0.004305834881961346, -0.09883370250463486, 0.03536948934197426, 0.0034318892285227776, 0.11122138053178787, -0.10920718312263489, -0.058959342539310455, 0.10603789985179901, -0.09977759420871735, -0.041270144283771515, 0.07228615880012512, -0.05243868753314018, 0.05399756878614426, 0.12966597080230713, 0.07704824954271317, 0.09160879999399185, -0.0738825798034668, -0.03476782515645027, -0.01279878057539463, 0.03623272478580475, 0.09734572470188141, 0.03739812970161438, -0.03146722912788391, -0.007890257984399796, 0.03125821053981781, -0.02048981562256813, -0.01968582347035408, 0.035462506115436554, -0.03083495981991291, -0.0050612520426511765, -0.018744349479675293, 0.00019364617764949799, -0.03672058507800102, -0.018150106072425842, -0.030721906572580338, -0.13716836273670197, 0.02437972091138363, 0.026188721880316734, -0.04252069443464279, 0.033014699816703796, -0.05959685146808624, 0.021574534475803375, -0.05870911478996277, -0.017537232488393784, -0.1666087806224823, -0.10901356488466263, 0.09603983908891678, -0.1149408370256424, 0.09753257036209106, 0.02167324721813202, -0.0005870405584573746, 0.03145601227879524, -0.046088140457868576, -0.05031955987215042, -0.1032804474234581, -0.014513062313199043, -0.047899261116981506, -0.09876805543899536, 0.020822517573833466, -0.03737790137529373, 0.023487258702516556, -0.07957702875137329, -0.0020187513437122107, 0.04670928791165352, 0.11824919283390045, 0.0996195375919342, -0.05943748727440834, 0.03408324718475342, -0.00808245874941349, -0.020020026713609695, -0.023475397378206253, -0.016357138752937317, 0.020846545696258545, -0.11689954996109009, 0.08077704906463623, -0.013741768896579742, -0.022507142275571823, 0.05742114782333374, 0.013078562915325165, -0.033446360379457474, -0.09704530239105225, -0.032724104821681976, -0.004837115295231342, -0.1035648062825203, -0.08168761432170868, 0.24603968858718872, 0.03746347874403, 0.07614672183990479, -0.10948360711336136, -0.038667842745780945, -0.027004670351743698, 0.002864317037165165, -0.00821862556040287, 0.018416371196508408, 0.05019286647439003, -0.0932643860578537, 0.005705304443836212, 0.0009808698669075966, 0.07103422284126282, 0.07483993470668793, -0.009405144490301609, -0.08272162079811096, 0.046128034591674805, -0.03903108090162277, -0.023712728172540665, 0.054488733410835266, -0.01197198498994112, 0.034528061747550964, 0.047674454748630524, 0.05954701453447342, 0.04392967000603676, -0.10767679661512375, 0.08499617874622345, 0.038344793021678925, -0.003608488477766514, -0.03521334007382393, -0.06191908195614815, 0.014492050744593143, 0.07900972664356232, 0.040215864777565, 0.04174811393022537, -0.022183964028954506, -0.05313201621174812, -0.09438794106245041, 0.12901240587234497, -0.0805913582444191, -0.2263195812702179, -0.1346416175365448, 0.04349694028496742, -0.061703406274318695, 0.0049807229079306126, -0.004040103405714035, -0.034035176038742065, -0.12701943516731262, -0.14471538364887238, 0.05412119999527931, 0.03538721799850464, -0.0448266863822937, -0.052254170179367065, 0.05340738594532013, 0.03769785165786743, -0.12636084854602814, -0.010194242000579834, -0.007293138653039932, 0.007836747914552689, -0.036611609160900116, 0.011653711088001728, 0.06806333363056183, 0.08374112099409103, 0.04793315753340721, -0.05476276949048042, -0.006449575535953045, 0.1495571881532669, -0.09102858603000641, 0.11769372969865799, 0.08674215525388718, -0.03556785359978676, 0.09791751205921173, 0.1253376603126526, 0.027697937563061714, -0.03992069140076637, 0.06126272305846214, 0.0610339492559433, 0.015400266274809837, -0.21384307742118835, -0.10433630645275116, -0.027126260101795197, -0.12646600604057312, 0.07604949921369553, 0.027564410120248795, -0.011712348088622093, 0.05076033994555473, -0.07963849604129791, 0.02547616884112358, 0.05420920252799988, 0.0449247732758522, 0.08840416371822357, 0.02820092812180519, 0.06944812834262848, -0.051312156021595, -0.04305412992835045, 0.08431259542703629, 0.040322694927453995, 0.2351057380437851, -0.05219978466629982, 0.10316191613674164, 0.037712469696998596, -0.021208060905337334, 0.034436654299497604, 0.09085293859243393, -0.04367467388510704, 0.030354980379343033, -0.05756661295890808, -0.050753992050886154, 0.024513674899935722, 0.11475314199924469, 0.010744921863079071, 0.0036050300113856792, -0.021030636504292488, -0.06511733680963516, 0.04630958288908005, 0.13472484052181244, 0.023829516023397446, -0.13390588760375977, -0.044763464480638504, 0.043437689542770386, -0.09638042747974396, -0.08924581110477448, 0.04215729981660843, 0.04074501991271973, -0.09981866925954819, 0.07811884582042694, 0.003857045667245984, 0.1319432109594345, -0.10117751359939575, -0.016033601015806198, -0.0031218044459819794, 0.05212049186229706, -0.04578825831413269, 0.09219641238451004, -0.07534611970186234, 0.08647094666957855, 0.00264004641212523, 0.06286544352769852, -0.07740218192338943, 0.05839234218001366, 0.03190266713500023, 0.02787429466843605, 0.046996377408504486, 0.010055774822831154, 0.04892180860042572, -0.010571926832199097, -0.1245054304599762, 0.04347798973321915, 0.037077657878398895, -0.06857415288686752, 0.10222356021404266, -0.011782808229327202, 0.02894846349954605, -0.01464915368705988, -0.08144208788871765, -0.17112493515014648, -0.18698516488075256, 0.015483122318983078, -0.05352506414055824, 0.05342905968427658, -0.07817748188972473, -0.06017196178436279, -0.02584819868206978, 0.12598702311515808, -0.07049070298671722, -0.0871240496635437, -0.13507875800132751, -0.04915079474449158, 0.10718148946762085, -0.042688194662332535, 0.0029186338651925325, -0.004929221235215664, 0.20694699883460999, -0.045714959502220154, -0.029860861599445343, 0.0514177568256855, -0.06672674417495728, -0.13230334222316742, -0.08555429428815842, 0.11501701921224594, 0.12119673192501068, 0.0396854430437088, 0.004738399758934975, -0.022551404312253, -0.030046602711081505, -0.0882229283452034, -0.02584679424762726, 0.13337081670761108, 0.02281319350004196, 0.06174927577376366, -0.08751443028450012, -0.044745072722435, -0.07203763723373413, 0.01928955316543579, -0.0008272787090390921, 0.06546665728092194, 0.016386594623327255, 0.10677433013916016, 0.11715882271528244, -0.06795333325862885, -0.1889394223690033, 0.05194452404975891, 0.06595522165298462, 0.009003303945064545, 0.040165822952985764, -0.18385492265224457, 0.1534060537815094, -0.0664810985326767, -0.037553731352090836, -0.006233674474060535, -0.10808639228343964, -0.10375472903251648, -0.01519017480313778, 0.019219305366277695, -0.02428176999092102, 0.000405977014452219, -0.05816150829195976, -0.02891852706670761, -0.07152619957923889, 0.12022340297698975, 0.009749280288815498, 0.022700684145092964, 0.03327115997672081, 0.058962371200323105, 0.04118068143725395, -0.02108103409409523, 0.05599905550479889, 0.0015993423294276, 0.04858008027076721, -0.010300828143954277, 0.07584939897060394, 0.08502675592899323, -0.04518852382898331, 0.09823648631572723, -0.029681984335184097, 0.029295343905687332, -0.08300349861383438, -0.06952360272407532, -0.04354165494441986, 0.09023428708314896, -0.019038625061511993, -0.06762935221195221, -0.09327816963195801, 0.05760681629180908, 0.11200318485498428, 0.0015139526221901178, 0.04132740944623947, -0.019981760531663895, -0.04254519194364548, 0.03916815295815468, 0.18566563725471497, -0.053809598088264465, -0.08614113181829453, -0.01760585978627205, -0.01414112001657486, 0.0826416015625, -0.13903295993804932, 0.05345622077584267, 0.07724149525165558, -0.03561161458492279, 0.1040952205657959, 0.005744965746998787, -0.15392455458641052, -0.0017417054623365402, 0.05868101865053177, -0.029208911582827568, -0.04336412623524666, -0.06732591986656189, 0.08059798181056976, -0.10789873450994492, -0.03064156137406826, 0.14993351697921753, -0.045555975288152695, -0.02773791365325451, -0.020076187327504158, 0.039016060531139374, 0.03795192018151283, 0.026594286784529686, 0.011406333185732365, 0.02585933543741703, -0.051161207258701324, 0.12862662971019745, 0.07875773310661316, -0.14179761707782745, -0.0044024051167070866, 0.005593547597527504, -0.09450291097164154, -0.007526954170316458, -0.017930734902620316, 0.14101500809192657, -0.0003637317568063736, -0.07623568177223206, -0.03036421537399292, -0.11440928280353546, 0.037527505308389664, 0.01859433390200138, 0.018948597833514214, 0.006096174940466881, -0.056164346635341644, 0.002324625849723816, -0.030646927654743195, 0.08051666617393494, 0.05444633215665817, 0.02578943595290184, -0.07288551330566406, -0.03012840449810028, 0.04644348844885826, -0.054002534598112106, -0.015702692791819572, -0.022652380168437958, -0.028254367411136627, -0.047188498079776764, -0.12564513087272644, 0.011020850390195847, -0.012231461703777313, 0.00191546231508255, -0.053624771535396576, -0.034220848232507706, -0.012504242360591888, 0.04747430980205536, -0.0463242344558239, -0.05486775562167168, -0.019726663827896118, 0.08477760851383209, -0.12187793105840683, 0.0345049723982811, 0.12876999378204346, -0.0920502245426178, 0.09347110986709595, -0.0518450029194355, -0.040881190448999405, 0.021268105134367943, -0.05759814754128456, 0.032265596091747284, 0.025547385215759277, 0.07563082873821259, -0.013012457638978958, -0.060775306075811386, 0.014782750979065895, 0.00022890337277203798, -0.043470874428749084, -0.055887095630168915, 0.035155847668647766, -0.0634625107049942, -0.005511738359928131, 0.026594676077365875, -0.07891257107257843, -0.035432368516922, 0.030701586976647377, 0.08361265808343887, 0.04455765336751938, 0.13869310915470123, 0.00917543563991785, 0.002487768419086933, -0.14058014750480652, 0.0014212103560566902, 0.045778822153806686, 0.011072914116084576, -0.05758851766586304, -0.05412546172738075, 0.024676969274878502, -0.009740264154970646, 0.23590229451656342, -0.005253700539469719, -0.009878048673272133, 0.04725199192762375, 0.017862826585769653, -0.04228496924042702, 0.030380453914403915, 0.03913800045847893, 0.02294335328042507, -0.008116600103676319, 0.01180572621524334, -0.06701481342315674, 0.0096976887434721, -0.032394155859947205, 0.05544591322541237, 0.10378424823284149, 0.13151639699935913, 0.050308793783187866, 0.024227473884820938, 0.014903751201927662, -0.14128535985946655, 0.051475755870342255, 0.03020656853914261, 0.04793780297040939, -0.02575693093240261, 0.133305162191391, 0.04887240752577782, -0.10324390977621078, 0.08826544880867004, 0.05218600481748581, -0.03659215196967125, -0.03118654526770115, -0.17182695865631104, -0.05080125853419304, -0.0564957857131958, 0.005379630718380213, -0.10575607419013977, -0.018343646079301834, 0.03486371040344238, -0.01996718905866146, -0.011430284939706326, 0.1399829387664795, -0.09660599380731583, -0.08944306522607803, 0.10046729445457458, 0.017292797565460205, 0.0014843214303255081, 0.10498452931642532, -0.016235489398241043, -0.0012371372431516647, 0.08760612457990646, 0.020515013486146927, 0.048242781311273575, -0.019317880272865295, 0.010787229984998703, 0.011891541071236134, -0.04690682142972946, -0.00044864416122436523, -0.0741659551858902, -0.03440190851688385, 0.08887679129838943, 0.048437077552080154, -0.019109874963760376, -0.017861729487776756, 0.23623788356781006, -0.04401158168911934, -0.007703398820012808, -0.13265961408615112, 0.03562900051474571, 0.005643528886139393, 0.01380356214940548, 0.04400699585676193, -0.10822993516921997, 0.023470595479011536, 0.16519922018051147, 0.07043930888175964, -0.08835683763027191, -0.013688729144632816, -0.010596671141684055, -0.014470264315605164, 0.003564384300261736, 0.059906817972660065, -0.007823570631444454, 0.18562962114810944, -0.018213670700788498, 0.0395871177315712, -0.04123222082853317, -0.0601646862924099, -0.008793623186647892, 0.1110927015542984, -0.02198035828769207, 0.0012767903972417116, -0.12200459837913513, 0.08749083429574966, -0.027840465307235718, -0.25753557682037354, -0.02958286926150322, -0.04484960436820984, -0.07913079112768173, -0.0023910312447696924, 0.007916543632745743, 0.01295609213411808, 0.06128957122564316, 0.055013999342918396, -0.025795869529247284, 0.11934822797775269, 0.005838335491716862, -0.012784678488969803, 0.01613081619143486, 0.05043552815914154, -0.08676151186227798, 0.18038734793663025, 0.04625634104013443, 0.03557984530925751, 0.06245538219809532, 0.03222902864217758, -0.0699787363409996, 0.038406237959861755, 0.020050805062055588, -0.010654130950570107, -0.004705403931438923, 0.07909154146909714, 0.02595987543463707, 0.08226019144058228, 0.09486745297908783, -0.04483289644122124, 0.06057929992675781, -0.03424510359764099, -0.09489963948726654, -0.09928446263074875, 0.04113113880157471, -0.09455622732639313, 0.1444767862558365, 0.12336750328540802, -0.023190662264823914, -0.012295261025428772, -0.034269265830516815, 0.026311244815587997, 0.007081163115799427, 0.0858551636338234, 0.0005764663219451904, -0.1239805594086647, -0.02272125892341137, 0.06463408470153809, 0.06192224845290184, -0.2182024121284485, -0.05060911551117897, -0.029315531253814697, -0.02187146432697773, -0.0018228543922305107, 0.07232337445020676, 0.07561759650707245, -0.016347549855709076, -0.026575639843940735, -0.03878201171755791, 0.020449437201023102, 0.0743328183889389, -0.07171757519245148, -0.0066031645983457565 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Bioformer-LitCovid-v1.3h This model is a fine-tuned version of [bioformers/bioformer-litcovid](https://huggingface.co/bioformers/bioformer-litcovid) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 0.8951 - Hamming loss: 0.0168 - F1 micro: 0.8565 - F1 macro: 0.3960 - F1 weighted: 0.8831 - F1 samples: 0.8789 - Precision micro: 0.7903 - Precision macro: 0.3221 - Precision weighted: 0.8426 - Precision samples: 0.8631 - Recall micro: 0.9348 - Recall macro: 0.6915 - Recall weighted: 0.9348 - Recall samples: 0.9435 - Roc Auc: 0.9604 - Accuracy: 0.6896 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0001 - train_batch_size: 16 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 3257 - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Hamming loss | F1 micro | F1 macro | F1 weighted | F1 samples | Precision micro | Precision macro | Precision weighted | Precision samples | Recall micro | Recall macro | Recall weighted | Recall samples | Roc Auc | Accuracy | |:-------------:|:-----:|:-----:|:---------------:|:------------:|:--------:|:--------:|:-----------:|:----------:|:---------------:|:---------------:|:------------------:|:-----------------:|:------------:|:------------:|:---------------:|:--------------:|:-------:|:--------:| | 1.2033 | 1.0 | 2272 | 0.5628 | 0.0616 | 0.6107 | 0.2167 | 0.7918 | 0.7257 | 0.4618 | 0.1789 | 0.7347 | 0.6771 | 0.9014 | 0.7310 | 0.9014 | 0.9194 | 0.9209 | 0.3870 | | 1.2127 | 2.0 | 4544 | 0.5062 | 0.0325 | 0.7555 | 0.2834 | 0.8357 | 0.8037 | 0.6337 | 0.2273 | 0.7680 | 0.7535 | 0.9353 | 0.7100 | 0.9353 | 0.9434 | 0.9523 | 0.4954 | | 0.96 | 3.0 | 6816 | 0.4943 | 0.0245 | 0.8043 | 0.3363 | 0.8608 | 0.8409 | 0.7043 | 0.2676 | 0.8069 | 0.8048 | 0.9372 | 0.7637 | 0.9372 | 0.9477 | 0.9575 | 0.5735 | | 0.5852 | 4.0 | 9088 | 0.7306 | 0.0195 | 0.8371 | 0.3860 | 0.8687 | 0.8624 | 0.7568 | 0.3083 | 0.8212 | 0.8378 | 0.9365 | 0.7232 | 0.9365 | 0.9459 | 0.9597 | 0.6410 | | 0.3454 | 5.0 | 11360 | 0.8951 | 0.0168 | 0.8565 | 0.3960 | 0.8831 | 0.8789 | 0.7903 | 0.3221 | 0.8426 | 0.8631 | 0.9348 | 0.6915 | 0.9348 | 0.9435 | 0.9604 | 0.6896 | ### Framework versions - Transformers 4.28.0 - Pytorch 2.0.0 - Datasets 2.1.0 - Tokenizers 0.13.3
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["accuracy"], "model-index": [{"name": "Bioformer-LitCovid-v1.3h", "results": []}]}
text-classification
sofia-todeschini/Bioformer-LitCovid-v1.3h
[ "transformers", "pytorch", "tensorboard", "bert", "text-classification", "generated_from_trainer", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-10T13:18:17+00:00
[]
[]
TAGS #transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
Bioformer-LitCovid-v1.3h ======================== This model is a fine-tuned version of bioformers/bioformer-litcovid on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 0.8951 * Hamming loss: 0.0168 * F1 micro: 0.8565 * F1 macro: 0.3960 * F1 weighted: 0.8831 * F1 samples: 0.8789 * Precision micro: 0.7903 * Precision macro: 0.3221 * Precision weighted: 0.8426 * Precision samples: 0.8631 * Recall micro: 0.9348 * Recall macro: 0.6915 * Recall weighted: 0.9348 * Recall samples: 0.9435 * Roc Auc: 0.9604 * Accuracy: 0.6896 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 0.0001 * train\_batch\_size: 16 * eval\_batch\_size: 32 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 3257 * num\_epochs: 5 ### Training results ### Framework versions * Transformers 4.28.0 * Pytorch 2.0.0 * Datasets 2.1.0 * Tokenizers 0.13.3
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 3257\n* num\\_epochs: 5", "### Training results", "### Framework versions\n\n\n* Transformers 4.28.0\n* Pytorch 2.0.0\n* Datasets 2.1.0\n* Tokenizers 0.13.3" ]
[ "TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 3257\n* num\\_epochs: 5", "### Training results", "### Framework versions\n\n\n* Transformers 4.28.0\n* Pytorch 2.0.0\n* Datasets 2.1.0\n* Tokenizers 0.13.3" ]
[ 55, 116, 4, 32 ]
[ "passage: TAGS\n#transformers #pytorch #tensorboard #bert #text-classification #generated_from_trainer #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.0001\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 3257\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.28.0\n* Pytorch 2.0.0\n* Datasets 2.1.0\n* Tokenizers 0.13.3" ]
[ -0.09670833498239517, 0.08500991761684418, -0.002238118089735508, 0.12142320722341537, 0.1677085906267166, 0.023672178387641907, 0.12345108389854431, 0.13642171025276184, -0.11173027753829956, 0.015072115696966648, 0.12397673726081848, 0.18004776537418365, 0.013137716799974442, 0.11272336542606354, -0.04984906315803528, -0.277706116437912, -0.014752859249711037, 0.031020360067486763, -0.09532874822616577, 0.13447515666484833, 0.08689349889755249, -0.12939536571502686, 0.080881267786026, 0.0025750307831913233, -0.18128599226474762, 0.018026411533355713, 0.01144411414861679, -0.06196530908346176, 0.15812602639198303, 0.019836733117699623, 0.1121697872877121, 0.017657967284321785, 0.09510629624128342, -0.19895333051681519, 0.012255439534783363, 0.05702982842922211, 0.010121774859726429, 0.08696041256189346, 0.06159387156367302, -0.002716902643442154, 0.13648344576358795, -0.08236861228942871, 0.069159597158432, 0.017816118896007538, -0.11686810851097107, -0.23685786128044128, -0.08325058966875076, 0.022518981248140335, 0.07202811539173126, 0.10583677887916565, 0.0023916056379675865, 0.13334324955940247, -0.08830229938030243, 0.10818835347890854, 0.24128277599811554, -0.2980668246746063, -0.062108226120471954, 0.02727946639060974, 0.027625633403658867, 0.07889493554830551, -0.11494012176990509, -0.01962531916797161, 0.033731717616319656, 0.04287134110927582, 0.1462785303592682, -0.03626246750354767, -0.10894723236560822, 0.021955739706754684, -0.13942620158195496, -0.030523093417286873, 0.11905231326818466, 0.02883743681013584, -0.02796974405646324, -0.05145886167883873, -0.0604344978928566, -0.1556084305047989, -0.03988533467054367, -0.012484384700655937, 0.05451570078730583, -0.035508524626493454, -0.07112977653741837, -0.021644916385412216, -0.1023155003786087, -0.07177326083183289, -0.04891352355480194, 0.1368459016084671, 0.041413117200136185, 0.010435719974339008, -0.030159631744027138, 0.11523798108100891, 0.03712838888168335, -0.14031445980072021, 0.021787188947200775, 0.034836865961551666, -0.01059651281684637, -0.04177292436361313, -0.06623725593090057, -0.04866112768650055, 0.006143699865788221, 0.1286960393190384, -0.0604746975004673, 0.06415560841560364, 0.025830265134572983, 0.03437696769833565, -0.10839337855577469, 0.19275179505348206, -0.036771900951862335, -0.02630901336669922, 0.0002615237608551979, 0.05959254503250122, 0.013791725039482117, -0.023803221061825752, -0.11846286058425903, 0.01976184919476509, 0.09992659091949463, 0.022756068035960197, -0.06438928097486496, 0.0753776952624321, -0.03831129148602486, -0.025169290602207184, 0.007479510735720396, -0.10478736460208893, 0.040146682411432266, 0.007963468320667744, -0.0910630002617836, -0.048465948551893234, 0.03261182829737663, 0.017894329503178596, -0.025444544851779938, 0.12938439846038818, -0.0813693031668663, 0.03623031824827194, -0.10303550213575363, -0.11388186365365982, 0.011883064173161983, -0.0778588280081749, 0.0229365024715662, -0.08984053134918213, -0.1781131476163864, -0.022291505709290504, 0.07026449590921402, -0.027022290974855423, -0.0529215931892395, -0.048813968896865845, -0.07567666471004486, 0.021547973155975342, -0.02398545667529106, 0.15636061131954193, -0.06600245088338852, 0.11442144960165024, 0.021701982244849205, 0.059638943523168564, -0.03320019319653511, 0.06201047822833061, -0.09517743438482285, 0.0017896296922117472, -0.17342273890972137, 0.04007674381136894, -0.05715435370802879, 0.059773728251457214, -0.09035494923591614, -0.10776917636394501, 0.03313044086098671, 0.004247469361871481, 0.07050272822380066, 0.08976487070322037, -0.1630443036556244, -0.07986502349376678, 0.1498880386352539, -0.0616094172000885, -0.10692526400089264, 0.10182265937328339, -0.06688982993364334, 0.04315480589866638, 0.06676577776670456, 0.17970024049282074, 0.0788734182715416, -0.07943272590637207, 0.03685011342167854, 0.002261590212583542, 0.05724853649735451, -0.05501789227128029, 0.06125396862626076, 0.00837706308811903, 0.016184885054826736, 0.02674206532537937, -0.04478059336543083, 0.06291177123785019, -0.08705544471740723, -0.09714054316282272, -0.03136957064270973, -0.09540297836065292, 0.04503242298960686, 0.0760294646024704, 0.08154283463954926, -0.09536110609769821, -0.08771788328886032, 0.07750225812196732, 0.0839928612112999, -0.06613661348819733, 0.038921058177948, -0.05434097722172737, 0.048419609665870667, -0.008414515294134617, -0.016421746462583542, -0.1934647262096405, -0.01889367401599884, 0.007930492982268333, -0.00600062683224678, 0.029477283358573914, 0.017968764528632164, 0.07785257697105408, 0.06589756906032562, -0.06354169547557831, -0.02584730088710785, -0.0250211451202631, -0.0014173469971865416, -0.13172678649425507, -0.21921442449092865, -0.03306511044502258, -0.020278869196772575, 0.12484516948461533, -0.2081780880689621, 0.03603031113743782, -0.012627032585442066, 0.06586558371782303, 0.018547670915722847, -0.007141378708183765, -0.046665191650390625, 0.0821487084031105, -0.03893600031733513, -0.054657042026519775, 0.08069366961717606, 0.008602161891758442, -0.08298593014478683, -0.03268589451909065, -0.11118905991315842, 0.17257285118103027, 0.135245218873024, -0.10994359850883484, -0.07922840863466263, 0.003113477025181055, -0.05570673570036888, -0.035903751850128174, -0.036509495228528976, 0.05064325034618378, 0.18522365391254425, -0.003459320869296789, 0.16164153814315796, -0.06456266343593597, -0.04956500604748726, 0.015937652438879013, -0.028092781081795692, 0.03270798176527023, 0.12800228595733643, 0.12532024085521698, -0.0646684467792511, 0.14035595953464508, 0.14286646246910095, -0.09591727703809738, 0.14203397929668427, -0.04195713251829147, -0.06810139119625092, -0.01175420731306076, -0.021797308698296547, -0.0003874995745718479, 0.09107266366481781, -0.12417565286159515, -0.0070624942891299725, 0.015933319926261902, 0.02372998744249344, 0.01870172657072544, -0.22907070815563202, -0.03227007016539574, 0.03513585403561592, -0.049195289611816406, -0.003940887283533812, -0.023566478863358498, -0.0010895958403125405, 0.1063714474439621, 0.00345138693228364, -0.09696803241968155, 0.031806979328393936, 0.0005377622437663376, -0.07552331686019897, 0.2134692221879959, -0.09172344952821732, -0.15742868185043335, -0.1162266805768013, -0.08570902794599533, -0.04744241386651993, 0.005202597472816706, 0.06494306027889252, -0.08864764124155045, -0.026191703975200653, -0.06505471467971802, 0.004192645661532879, -0.0015329885063692927, 0.02847796306014061, -0.005896960850805044, -0.001589699531905353, 0.06463396549224854, -0.11172334104776382, -0.00922057032585144, -0.05474569648504257, -0.06510034948587418, 0.043320704251527786, 0.054915901273489, 0.10686878859996796, 0.1570957750082016, -0.013605117797851562, 0.01255050953477621, -0.03575754910707474, 0.20780684053897858, -0.07350116968154907, -0.019319623708724976, 0.1370648890733719, -0.013220051303505898, 0.05633719637989998, 0.11067712306976318, 0.07001864165067673, -0.09022139012813568, 0.01629200205206871, 0.041151318699121475, -0.033537644892930984, -0.22446560859680176, -0.03953114151954651, -0.053882863372564316, -0.013413134030997753, 0.09454318135976791, 0.031804684549570084, 0.0485931932926178, 0.04971446841955185, 0.05053838714957237, 0.06250827014446259, -0.02132953517138958, 0.057051654905080795, 0.11451184004545212, 0.04427296295762062, 0.13188880681991577, -0.041301071643829346, -0.06114659458398819, 0.03812536224722862, -0.010529710911214352, 0.21310651302337646, -0.0018221630016341805, 0.12173505127429962, 0.05007234215736389, 0.17417046427726746, -0.009066924452781677, 0.07745376974344254, -0.002031597075983882, -0.036868318915367126, -0.016275018453598022, -0.04024916887283325, -0.03948219120502472, 0.025925686582922935, -0.054804541170597076, 0.042016591876745224, -0.11958272755146027, -0.003517631907016039, 0.04944441840052605, 0.28112754225730896, 0.027815040200948715, -0.3133573830127716, -0.08196369558572769, 0.0035465664695948362, -0.04776793718338013, -0.018891623243689537, 0.027774807065725327, 0.09252546727657318, -0.1024303287267685, 0.03718430921435356, -0.07772215455770493, 0.09141620248556137, -0.04761454463005066, 0.04821038618683815, 0.08202758431434631, 0.08343152701854706, 0.007800641935318708, 0.07719191163778305, -0.31179922819137573, 0.27876007556915283, 0.0028184300754219294, 0.06457255035638809, -0.07899844646453857, 0.008549954742193222, 0.051283497363328934, 0.059877216815948486, 0.06143805384635925, -0.02039489336311817, -0.05627569556236267, -0.19468910992145538, -0.055112194269895554, 0.025427399203181267, 0.09723369777202606, -0.03751648962497711, 0.09388928860425949, -0.0398191437125206, -0.0025713632348924875, 0.07415760308504105, -0.02800682559609413, -0.046713415533304214, -0.09836903214454651, -0.0015346388099715114, 0.01414667721837759, -0.034469909965991974, -0.0652056410908699, -0.12039823085069656, -0.10804621130228043, 0.1628853976726532, -0.02451455220580101, -0.030452514067292213, -0.12394808232784271, 0.08054770529270172, 0.0697532668709755, -0.0889279842376709, 0.03882884979248047, 0.01367486733943224, 0.060402724891901016, 0.033517248928546906, -0.07867954671382904, 0.12084978073835373, -0.06519994139671326, -0.1733543425798416, -0.05491755157709122, 0.10078974068164825, 0.03551021218299866, 0.06300730258226395, -0.02257806994020939, 0.021765215322375298, -0.028777500614523888, -0.08758381009101868, 0.025667816400527954, -0.02461058646440506, 0.07275860011577606, 0.011048729531466961, -0.06466671824455261, 0.02109670639038086, -0.058612193912267685, -0.03781970962882042, 0.18944144248962402, 0.23723526298999786, -0.09789086133241653, 0.05037972703576088, 0.04368429630994797, -0.07473617792129517, -0.20966015756130219, 0.035740215331315994, 0.06388843804597855, -0.0045532784424722195, 0.059075865894556046, -0.201873779296875, 0.11923254281282425, 0.08608006685972214, -0.018838196992874146, 0.10910901427268982, -0.32492977380752563, -0.13694456219673157, 0.12378031760454178, 0.1489715725183487, 0.09036299586296082, -0.15348944067955017, -0.01919085904955864, -0.018887486308813095, -0.10211508721113205, 0.10079465061426163, -0.08313969522714615, 0.1280602514743805, -0.02477892115712166, 0.08100318908691406, 0.00554688973352313, -0.05803554132580757, 0.10892133414745331, 0.003673538565635681, 0.10429881513118744, -0.06361441314220428, -0.02001391351222992, 0.03115805983543396, -0.04554956033825874, 0.012699661776423454, -0.10239146649837494, 0.02789711393415928, -0.09323034435510635, -0.021995816379785538, -0.08547189086675644, 0.03621520474553108, -0.03857260197401047, -0.06670767068862915, -0.026768427342176437, 0.031764667481184006, 0.056595008820295334, -0.01153726875782013, 0.1476919800043106, 0.007476709317415953, 0.14964815974235535, 0.12401774525642395, 0.07430122792720795, -0.05596422031521797, -0.06702416390180588, -0.021911146119236946, -0.012443721294403076, 0.05805463716387749, -0.1481136828660965, 0.03530772775411606, 0.14317260682582855, 0.018558461219072342, 0.13669417798519135, 0.07919161021709442, -0.01881030946969986, 0.0038904689718037844, 0.05866341292858124, -0.16639237105846405, -0.07012273371219635, -0.004125781357288361, -0.05986282229423523, -0.0966244712471962, 0.045272935181856155, 0.10544528067111969, -0.06186142563819885, -0.009352694265544415, -0.0022361029405146837, 0.015758082270622253, -0.04616451635956764, 0.20640185475349426, 0.057664137333631516, 0.05063093081116676, -0.10422354191541672, 0.08096685260534286, 0.055148083716630936, -0.08705098181962967, 0.02074979618191719, 0.0976269468665123, -0.08119076490402222, -0.04410891979932785, 0.09973396360874176, 0.17503294348716736, -0.05241205170750618, -0.04557162523269653, -0.14689363539218903, -0.12134665250778198, 0.08308151364326477, 0.16006232798099518, 0.10409175604581833, 0.007655350957065821, -0.06852839142084122, 0.02274596318602562, -0.11739777028560638, 0.09670418500900269, 0.03855695202946663, 0.06766046583652496, -0.1396394520998001, 0.16279077529907227, 0.01670992374420166, 0.04414468631148338, -0.0218979362398386, 0.023834511637687683, -0.11053470522165298, 0.015314685180783272, -0.12385337054729462, -0.03126109391450882, -0.016042513772845268, 0.009366092272102833, -0.005230962298810482, -0.05665344372391701, -0.057563308626413345, 0.019405722618103027, -0.12530381977558136, -0.029696762561798096, 0.021028853952884674, 0.05302366241812706, -0.1283373087644577, -0.04764042794704437, 0.025667376816272736, -0.07065627723932266, 0.06959309428930283, 0.04015238955616951, 0.010558386333286762, 0.06902800500392914, -0.12898756563663483, 0.001775412354618311, 0.05715642496943474, 0.011923783458769321, 0.0589388944208622, -0.09295317530632019, -0.008218478411436081, -0.00624859007075429, 0.06314947456121445, 0.018709523603320122, 0.0875861719250679, -0.14030560851097107, 0.012009386904537678, -0.029473667964339256, -0.09057767689228058, -0.06366011500358582, 0.03928850218653679, 0.07355803996324539, 0.011133487336337566, 0.19689646363258362, -0.08823712170124054, 0.04093771427869797, -0.21388861536979675, 0.0006018663407303393, -0.011957100592553616, -0.12118709087371826, -0.12503157556056976, -0.07169464975595474, 0.07365258038043976, -0.06260746717453003, 0.122174933552742, 0.04092025384306908, 0.056623127311468124, 0.03256789594888687, -0.029800688847899437, 0.021325230598449707, 0.02712358348071575, 0.1877400279045105, 0.03270133584737778, -0.03203082084655762, 0.08237447589635849, 0.0552576407790184, 0.09487966448068619, 0.11876322329044342, 0.22080650925636292, 0.14088454842567444, 0.0038658124394714832, 0.09133066982030869, 0.030273819342255592, -0.07094398885965347, -0.16881263256072998, 0.03697678819298744, -0.05057529732584953, 0.1210395023226738, -0.03685658052563667, 0.21570467948913574, 0.06249326467514038, -0.17395006120204926, 0.053517770022153854, -0.059838708490133286, -0.08967185765504837, -0.12311797589063644, -0.038763757795095444, -0.08203539252281189, -0.1490727812051773, 0.0004597009683493525, -0.10532142966985703, 0.031310565769672394, 0.10438825190067291, 0.007132707163691521, -0.01682320423424244, 0.1482296586036682, 0.015920544043183327, 0.0229815561324358, 0.054758235812187195, 0.0047352067194879055, -0.02919471263885498, -0.09880146384239197, -0.07335226982831955, -0.02388618141412735, -0.015138518065214157, 0.02755645662546158, -0.06414913386106491, -0.07502327114343643, 0.045489683747291565, -0.025096355006098747, -0.09384610503911972, 0.021684007719159126, 0.014089544303715229, 0.07751926779747009, 0.05497461557388306, 0.008039062842726707, 0.008311750367283821, -0.0093496423214674, 0.24270319938659668, -0.08868268877267838, -0.07913655787706375, -0.0919538140296936, 0.2763839662075043, 0.05053325742483139, -0.014179677702486515, 0.030182480812072754, -0.06299072504043579, -0.01146298460662365, 0.2558407485485077, 0.20967158675193787, -0.09737865626811981, -0.012044725939631462, 0.010327010415494442, -0.005632862914353609, -0.018150297924876213, 0.12943534553050995, 0.14093154668807983, 0.05588486045598984, -0.10147662460803986, -0.05338742583990097, -0.042157139629125595, -0.024152847006917, -0.04730993136763573, 0.07960733771324158, 0.04105963930487633, 0.003021112410351634, -0.03189195320010185, 0.05140986666083336, -0.05693194642663002, -0.12759631872177124, 0.05793289095163345, -0.20834881067276, -0.16782262921333313, -0.013743191957473755, 0.11993398517370224, -0.00030859149410389364, 0.058098968118429184, -0.009129048325121403, -0.0029721036553382874, 0.07061715424060822, -0.01863889768719673, -0.08218575268983841, -0.08265887945890427, 0.10423185676336288, -0.13698288798332214, 0.2076919823884964, -0.0521501861512661, 0.04669521749019623, 0.12413016706705093, 0.06933815777301788, -0.06227704882621765, 0.062221087515354156, 0.04928836598992348, -0.09983565658330917, 0.0312805138528347, 0.093410424888134, -0.041211895644664764, 0.06762871146202087, 0.03818570077419281, -0.11931150406599045, 0.027332987636327744, -0.07889610528945923, -0.06854233890771866, -0.029698947444558144, -0.04407627508044243, -0.05223887413740158, 0.12436557561159134, 0.2183276265859604, -0.023851940408349037, 0.016750957816839218, -0.08318151533603668, 0.004030149895697832, 0.04332992807030678, 0.02157668024301529, -0.07925401628017426, -0.23036830127239227, 0.0168943889439106, 0.04324500635266304, -0.015891479328274727, -0.23376329243183136, -0.09860867261886597, 0.016644280403852463, -0.06362754106521606, -0.10815750807523727, 0.0880720391869545, 0.07482059299945831, 0.04272027313709259, -0.05125916749238968, -0.07813692092895508, -0.07595013827085495, 0.16029095649719238, -0.16861353814601898, -0.0814603939652443 ]
null
null
transformers
<div align="center"> <h1> MiniCPM </h1> </div> <p align="center"> <a href="https://shengdinghu.notion.site/MiniCPM-c805a17c5c8046398914e47f0542095a?pvs=4" target="_blank">MiniCPM 技术报告</a><a href="https://shengdinghu.notion.site/MiniCPM-Unveiling-the-Potential-of-End-side-Large-Language-Models-d4d3a8c426424654a4e80e42a711cb20?pvs=4" target="_blank"> Technical Report</a> | <a href="https://github.com/OpenBMB/OmniLMM/" target="_blank">OmniLMM 多模态模型 Multi-modal Model</a> | <a href="https://luca.cn/" target="_blank">CPM-C 千亿模型试用 ~100B Model Trial </a> </p> MiniCPM 是面壁与清华大学自然语言处理实验室共同开源的系列端侧语言大模型,主体语言模型 MiniCPM-2B 仅有 24亿(2.4B)的非词嵌入参数量。 - 经过 SFT 后,MiniCPM 在公开综合性评测集上,MiniCPM 与 Mistral-7B相近(中文、数学、代码能力更优),整体性能超越 Llama2-13B、MPT-30B、Falcon-40B 等模型。 - 经过 DPO 后,MiniCPM 在当前最接近用户体感的评测集 MTBench上,MiniCPM-2B 也超越了 Llama2-70B-Chat、Vicuna-33B、Mistral-7B-Instruct-v0.1、Zephyr-7B-alpha 等众多代表性开源大模型。 - 以 MiniCPM-2B 为基础构建端侧多模态大模型 MiniCPM-V,整体性能在同规模模型中实现最佳,超越基于 Phi-2 构建的现有多模态大模型,在部分评测集上达到与 9.6B Qwen-VL-Chat 相当甚至更好的性能。 - 经过 Int4 量化后,MiniCPM 可在手机上进行部署推理,流式输出速度略高于人类说话速度。MiniCPM-V 也首次跑通了多模态大模型在手机上的部署。 - 一张1080/2080可高效参数微调,一张3090/4090可全参数微调,一台机器可持续训练 MiniCPM,二次开发成本较低。 我们将完全开源MiniCPM-2B的模型参数供学术研究和有限商用,以及训练过程中的所有Checkpoint和大部分非专有数据供模型机理研究。 - 基于MiniCPM-2B的指令微调与人类偏好对**MiniCPM-2B-SFT/DPO。** - 基于MiniCPM-2B的多模态模型**MiniCPM-V**,能力超越基于Phi-2的同参数级别多模态模型**。** - MiniCPM-2B-SFT/DPO的Int4量化版**MiniCPM-2B-SFT/DPO-Int4。** - 基于MLC-LLM、LLMFarm开发的MiniCPM手机端程序,**文本及多模态模型均可在手机端进行推理。** MiniCPM is an End-Size LLM developed by ModelBest Inc. and TsinghuaNLP, with only 2.4B parameters excluding embeddings. - MiniCPM has very close performance compared with Mistral-7B on open-sourced general benchmarks with better ability on Chinese, Mathmetics and Coding after SFT. The overall performance exceeds Llama2-13B, MPT-30B, Falcon-40B, etc. - After DPO, MiniCPM outperforms Llama2-70B-Chat, Vicuna-33B, Mistral-7B-Instruct-v0.1, Zephyr-7B-alpha, etc. on MTBench. - MiniCPM-V, based on MiniCPM-2B, achieves the best overall performance among multimodel models of the same scale, surpassing existing multimodal large models built on Phi-2 and achieving performance comparable to or even better than 9.6B Qwen-VL-Chat on some tasks. - MiniCPM can be deployed and infer on smartphones, and the speed of streaming output is relatively higher than the verbal speed of human. MiniCPM-V is the first multi-modal models that can be deployed on smartphones. - The cost of developing based on MiniCPM is low. Parameter efficient finetuning can be conducted with a single 1080/2080 GPU and full parameter finetuning can be conducted with a 3090/4090 GPU. We release all model parameters for research and limited commercial use. We also release all the checkpoint during training and most public training data for research on model mechanism. - SFT and DPO version based on MiniCPM-2B and human preference: **MiniCPM-2B-SFT/DPO** - The multi-modal model **MiniCPM-V** based on MiniCPM-2B, which outperforms models with similar size, i.e., Phi-2 - The INT4 quantized version **MiniCPM-2B-SFT/DPO-Int4** based on MiniCPM-2B-SFT/DPO - Mobile phone application based on MLC-LLM and LLMFarm. Both language model and multimodel model can conduct inference on smartphones. ### 评测结果 Evaluation Results 详细的评测结果位于[github仓库](https://github.com/OpenBMB/MiniCPM?tab=readme-ov-file#%E8%AF%84%E6%B5%8B%E7%BB%93%E6%9E%9C) Detailed evaluation results are in [github repo](https://github.com/OpenBMB/MiniCPM/blob/main/README-en.md#evaluation-results) 注意:我们发现使用Huggingface生成质量略差于vLLM,因此推荐使用vLLM进行测试。我们正在排查原因。 Notice: We discovered that the quality of Huggingface generation is slightly lower than vLLM, thus benchmarking using vLLM is recommended. We are investigating the cause now. ### 局限性 Limitations - 受限于模型规模,模型可能出现幻觉性问题。其中由于DPO模型生成的回复内容更长,更容易出现幻觉。我们也将持续进行MiniCPM模型的迭代改进; - 为了保证在学术研究用途上模型的通用性,我们未对模型进行任何身份认同训练。同时由于我们用ShareGPT开源语料作为部分训练数据,模型可能会输出类似GPT系列模型的身份认同信息; - 受限于模型规模,模型的输出受到提示词(prompt)的影响较大,可能多次尝试产生不一致的结果; - 受限于模型容量,模型的知识记忆较不准确,后续我们将结合RAG方法来增强模型的知识记忆能力。 - Due to limitations in model size, the model may experience hallucinatory issues. As DPO model tend to generate longer response, hallucinations are more likely to occur. We will also continue to iterate and improve the MiniCPM model. - To ensure the universality of the model for academic research purposes, we did not conduct any identity training on the model. Meanwhile, as we use ShareGPT open-source corpus as part of the training data, the model may output identity information similar to the GPT series models. - Due to the limitation of model size, the output of the model is greatly influenced by prompt words, which may result in inconsistent results from multiple attempts. - Due to limited model capacity, the model's knowledge memory is not accurate. In the future, we will combine the RAG method to enhance the model's knowledge memory ability. ## 模型下载 Download | HuggingFace | ModelScope | WiseModel | |-------------|------------|-----------| |[sft-bf16](https://huggingface.co/openbmb/MiniCPM-2B-sft-bf16)|[sft-bf16](https://modelscope.cn/models/OpenBMB/miniCPM-bf16)|[sft-bf16](https://wisemodel.cn/models/OpenBMB/miniCPM-bf16) |[sft-fp32](https://huggingface.co/openbmb/MiniCPM-2B-sft-fp32)|[sft-fp32](https://modelscope.cn/models/OpenBMB/MiniCPM-2B-sft-fp32)|[sft-fp32](https://wisemodel.cn/models/OpenBMB/miniCPM-dpo-fp32) |[dpo-bf16](https://huggingface.co/openbmb/MiniCPM-2B-dpo-bf16)|[dpo-bf16](https://modelscope.cn/models/OpenBMB/MiniCPM-2B-dpo-bf16/summary)|[dpo-bf16](https://wisemodel.cn/models/OpenBMB/MiniCPM-2B-dpo-bf16) |[dpo-fp16](https://huggingface.co/openbmb/MiniCPM-2B-dpo-fp16)|[dpo-fp16](https://modelscope.cn/models/OpenBMB/MiniCPM-2B-dpo-fp16/)|[dpo-fp16](https://wisemodel.cn/models/OpenBMB/MiniCPM-2B-dpo-fp16) |[dpo-fp32](https://huggingface.co/openbmb/MiniCPM-2B-dpo-fp32)|[dpo-fp32](https://modelscope.cn/models/OpenBMB/MiniCPM-2B-dpo-fp32)|[dpo-fp32](https://wisemodel.cn/models/OpenBMB/miniCPM-dpo-fp32) ## 模型使用 Usage * 安装`transformers>=4.36.0`以及`accelerate`后,运行以下代码 * 注意:需要在`from_pretrained`中明确指明模型的数据类型,否则会引起较大计算误差 * Run the following code after install `transformers>=4.36.0` and `accelerate` * Warning: It is necessary to specify the data type of the model clearly in 'from_pretrained', otherwise large calculation errors will be caused ```python from transformers import AutoModelForCausalLM, AutoTokenizer import torch torch.manual_seed(0) path = 'openbmb/MiniCPM-2B-dpo-bf16' tokenizer = AutoTokenizer.from_pretrained(path) model = AutoModelForCausalLM.from_pretrained(path, torch_dtype=torch.bfloat16, device_map='cuda', trust_remote_code=True) responds, history = model.chat(tokenizer, "山东省最高的山是哪座山, 它比黄山高还是矮?差距多少?", temperature=0.8, top_p=0.8) print(responds) ``` * 期望输出 Expected Output ```shell 山东省最高的山是泰山,海拔1545米。 相对于黄山(海拔1864米),泰山海拔较低,相差约319米。 ``` ## 开源协议 LICENSE #### 模型协议 Model LICENSE * 本仓库中代码依照 [Apache-2.0](https://github.com/OpenBMB/MiniCPM/blob/main/LICENSE) 协议开源 * MiniCPM 模型权重的使用则需要遵循 [“通用模型许可协议-来源说明-宣传限制-商业授权”](https://github.com/OpenBMB/General-Model-License/blob/main/%E9%80%9A%E7%94%A8%E6%A8%A1%E5%9E%8B%E8%AE%B8%E5%8F%AF%E5%8D%8F%E8%AE%AE-%E6%9D%A5%E6%BA%90%E8%AF%B4%E6%98%8E-%E5%AE%A3%E4%BC%A0%E9%99%90%E5%88%B6-%E5%95%86%E4%B8%9A%E6%8E%88%E6%9D%83.md)。 * MiniCPM 模型权重对学术研究完全开放。 * 如需将模型用于商业用途,请联系[email protected]来获取书面授权,在登记后亦允许免费商业使用。 * This repository is released under the [Apache-2.0](https://github.com/OpenBMB/MiniCPM/blob/main/LICENSE) License. * The usage of MiniCPM model weights must strictly follow [the General Model License (GML)](https://github.com/OpenBMB/General-Model-License/blob/main/%E9%80%9A%E7%94%A8%E6%A8%A1%E5%9E%8B%E8%AE%B8%E5%8F%AF%E5%8D%8F%E8%AE%AE-%E6%9D%A5%E6%BA%90%E8%AF%B4%E6%98%8E-%E5%AE%A3%E4%BC%A0%E9%99%90%E5%88%B6-%E5%95%86%E4%B8%9A%E6%8E%88%E6%9D%83.md). * The models and weights of MiniCPM are completely free for academic research. * If you intend to utilize the model for commercial purposes, please reach out to [email protected] to obtain the certificate of authorization. #### 声明 Statement * 作为一个语言模型,MiniCPM 通过学习大量的文本来生成内容,但它无法理解、表达个人观点或价值判断,它所输出的任何内容都不代表模型开发者的观点和立场。 * 因此用户在使用 MiniCPM 生成的内容时,应自行负责对其进行评估和验证。 * 如果由于使用 MinCPM 开源模型而导致的任何问题,包括但不限于数据安全问题、公共舆论风险,或模型被误导、滥用、传播或不当利用所带来的任何风险和问题,我们将不承担任何责任。 * As a language model, MiniCPM generates content by learning from a vast amount of text. * However, it does not possess the ability to comprehend or express personal opinions or value judgments. * Any content generated by MiniCPM does not represent the viewpoints or positions of the model developers. * Therefore, when using content generated by MiniCPM, users should take full responsibility for evaluating and verifying it on their own. <p id="8"></p> ## 工作引用 Citation * 如果觉得MiniCPM有助于您的工作,请考虑引用下列[技术报告](https://shengdinghu.notion.site/MiniCPM-c805a17c5c8046398914e47f0542095a?pvs=4) * Please cite our [techinical report](https://shengdinghu.notion.site/MiniCPM-Unveiling-the-Potential-of-End-side-Large-Language-Models-d4d3a8c426424654a4e80e42a711cb20?pvs=4) if you find our work valuable. ``` @inproceedings{minicpm2024, title={MiniCPM:Unveiling the Potential of End-side Large Language Models}, booktitle={OpenBMB Blog}, year={2024} } ```
{"language": ["en", "zh"], "tags": ["MiniCPM", "ModelBest", "THUNLP"]}
null
jncraton/MiniCPM-2B-dpo-bf16-llama-format-ct2-int8
[ "transformers", "MiniCPM", "ModelBest", "THUNLP", "en", "zh", "endpoints_compatible", "region:us" ]
2024-02-10T13:23:32+00:00
[]
[ "en", "zh" ]
TAGS #transformers #MiniCPM #ModelBest #THUNLP #en #zh #endpoints_compatible #region-us
MiniCPM ========= [MiniCPM 技术报告](URL target=) [Technical Report](URL target=) | [OmniLMM 多模态模型 Multi-modal Model](URL target=) | [CPM-C 千亿模型试用 ~100B Model Trial](URL target=) MiniCPM 是面壁与清华大学自然语言处理实验室共同开源的系列端侧语言大模型,主体语言模型 MiniCPM-2B 仅有 24亿(2.4B)的非词嵌入参数量。 * 经过 SFT 后,MiniCPM 在公开综合性评测集上,MiniCPM 与 Mistral-7B相近(中文、数学、代码能力更优),整体性能超越 Llama2-13B、MPT-30B、Falcon-40B 等模型。 * 经过 DPO 后,MiniCPM 在当前最接近用户体感的评测集 MTBench上,MiniCPM-2B 也超越了 Llama2-70B-Chat、Vicuna-33B、Mistral-7B-Instruct-v0.1、Zephyr-7B-alpha 等众多代表性开源大模型。 * 以 MiniCPM-2B 为基础构建端侧多模态大模型 MiniCPM-V,整体性能在同规模模型中实现最佳,超越基于 Phi-2 构建的现有多模态大模型,在部分评测集上达到与 9.6B Qwen-VL-Chat 相当甚至更好的性能。 * 经过 Int4 量化后,MiniCPM 可在手机上进行部署推理,流式输出速度略高于人类说话速度。MiniCPM-V 也首次跑通了多模态大模型在手机上的部署。 * 一张1080/2080可高效参数微调,一张3090/4090可全参数微调,一台机器可持续训练 MiniCPM,二次开发成本较低。 我们将完全开源MiniCPM-2B的模型参数供学术研究和有限商用,以及训练过程中的所有Checkpoint和大部分非专有数据供模型机理研究。 * 基于MiniCPM-2B的指令微调与人类偏好对MiniCPM-2B-SFT/DPO。 * 基于MiniCPM-2B的多模态模型MiniCPM-V,能力超越基于Phi-2的同参数级别多模态模型。 * MiniCPM-2B-SFT/DPO的Int4量化版MiniCPM-2B-SFT/DPO-Int4。 * 基于MLC-LLM、LLMFarm开发的MiniCPM手机端程序,文本及多模态模型均可在手机端进行推理。 MiniCPM is an End-Size LLM developed by ModelBest Inc. and TsinghuaNLP, with only 2.4B parameters excluding embeddings. * MiniCPM has very close performance compared with Mistral-7B on open-sourced general benchmarks with better ability on Chinese, Mathmetics and Coding after SFT. The overall performance exceeds Llama2-13B, MPT-30B, Falcon-40B, etc. * After DPO, MiniCPM outperforms Llama2-70B-Chat, Vicuna-33B, Mistral-7B-Instruct-v0.1, Zephyr-7B-alpha, etc. on MTBench. * MiniCPM-V, based on MiniCPM-2B, achieves the best overall performance among multimodel models of the same scale, surpassing existing multimodal large models built on Phi-2 and achieving performance comparable to or even better than 9.6B Qwen-VL-Chat on some tasks. * MiniCPM can be deployed and infer on smartphones, and the speed of streaming output is relatively higher than the verbal speed of human. MiniCPM-V is the first multi-modal models that can be deployed on smartphones. * The cost of developing based on MiniCPM is low. Parameter efficient finetuning can be conducted with a single 1080/2080 GPU and full parameter finetuning can be conducted with a 3090/4090 GPU. We release all model parameters for research and limited commercial use. We also release all the checkpoint during training and most public training data for research on model mechanism. * SFT and DPO version based on MiniCPM-2B and human preference: MiniCPM-2B-SFT/DPO * The multi-modal model MiniCPM-V based on MiniCPM-2B, which outperforms models with similar size, i.e., Phi-2 * The INT4 quantized version MiniCPM-2B-SFT/DPO-Int4 based on MiniCPM-2B-SFT/DPO * Mobile phone application based on MLC-LLM and LLMFarm. Both language model and multimodel model can conduct inference on smartphones. ### 评测结果 Evaluation Results 详细的评测结果位于github仓库 Detailed evaluation results are in github repo 注意:我们发现使用Huggingface生成质量略差于vLLM,因此推荐使用vLLM进行测试。我们正在排查原因。 Notice: We discovered that the quality of Huggingface generation is slightly lower than vLLM, thus benchmarking using vLLM is recommended. We are investigating the cause now. ### 局限性 Limitations * 受限于模型规模,模型可能出现幻觉性问题。其中由于DPO模型生成的回复内容更长,更容易出现幻觉。我们也将持续进行MiniCPM模型的迭代改进; * 为了保证在学术研究用途上模型的通用性,我们未对模型进行任何身份认同训练。同时由于我们用ShareGPT开源语料作为部分训练数据,模型可能会输出类似GPT系列模型的身份认同信息; * 受限于模型规模,模型的输出受到提示词(prompt)的影响较大,可能多次尝试产生不一致的结果; * 受限于模型容量,模型的知识记忆较不准确,后续我们将结合RAG方法来增强模型的知识记忆能力。 * Due to limitations in model size, the model may experience hallucinatory issues. As DPO model tend to generate longer response, hallucinations are more likely to occur. We will also continue to iterate and improve the MiniCPM model. * To ensure the universality of the model for academic research purposes, we did not conduct any identity training on the model. Meanwhile, as we use ShareGPT open-source corpus as part of the training data, the model may output identity information similar to the GPT series models. * Due to the limitation of model size, the output of the model is greatly influenced by prompt words, which may result in inconsistent results from multiple attempts. * Due to limited model capacity, the model's knowledge memory is not accurate. In the future, we will combine the RAG method to enhance the model's knowledge memory ability. 模型下载 Download ------------- HuggingFace: sft-bf16, ModelScope: sft-bf16, WiseModel: sft-bf16 HuggingFace: sft-fp32, ModelScope: sft-fp32, WiseModel: sft-fp32 HuggingFace: dpo-bf16, ModelScope: dpo-bf16, WiseModel: dpo-bf16 HuggingFace: dpo-fp16, ModelScope: dpo-fp16, WiseModel: dpo-fp16 HuggingFace: dpo-fp32, ModelScope: dpo-fp32, WiseModel: dpo-fp32 模型使用 Usage ---------- * 安装'transformers>=4.36.0'以及'accelerate'后,运行以下代码 * 注意:需要在'from\_pretrained'中明确指明模型的数据类型,否则会引起较大计算误差 * Run the following code after install 'transformers>=4.36.0' and 'accelerate' * Warning: It is necessary to specify the data type of the model clearly in 'from\_pretrained', otherwise large calculation errors will be caused * 期望输出 Expected Output 开源协议 LICENSE ------------ #### 模型协议 Model LICENSE * 本仓库中代码依照 Apache-2.0 协议开源 * MiniCPM 模型权重的使用则需要遵循 “通用模型许可协议-来源说明-宣传限制-商业授权”。 * MiniCPM 模型权重对学术研究完全开放。 * 如需将模型用于商业用途,请联系[email protected]来获取书面授权,在登记后亦允许免费商业使用。 * This repository is released under the Apache-2.0 License. * The usage of MiniCPM model weights must strictly follow the General Model License (GML). * The models and weights of MiniCPM are completely free for academic research. * If you intend to utilize the model for commercial purposes, please reach out to cpm@URL to obtain the certificate of authorization. #### 声明 Statement * 作为一个语言模型,MiniCPM 通过学习大量的文本来生成内容,但它无法理解、表达个人观点或价值判断,它所输出的任何内容都不代表模型开发者的观点和立场。 * 因此用户在使用 MiniCPM 生成的内容时,应自行负责对其进行评估和验证。 * 如果由于使用 MinCPM 开源模型而导致的任何问题,包括但不限于数据安全问题、公共舆论风险,或模型被误导、滥用、传播或不当利用所带来的任何风险和问题,我们将不承担任何责任。 * As a language model, MiniCPM generates content by learning from a vast amount of text. * However, it does not possess the ability to comprehend or express personal opinions or value judgments. * Any content generated by MiniCPM does not represent the viewpoints or positions of the model developers. * Therefore, when using content generated by MiniCPM, users should take full responsibility for evaluating and verifying it on their own. 工作引用 Citation ------------- * 如果觉得MiniCPM有助于您的工作,请考虑引用下列技术报告 * Please cite our techinical report if you find our work valuable.
[ "### 评测结果 Evaluation Results\n\n\n详细的评测结果位于github仓库\n\n\nDetailed evaluation results are in github repo\n\n\n注意:我们发现使用Huggingface生成质量略差于vLLM,因此推荐使用vLLM进行测试。我们正在排查原因。\n\n\nNotice: We discovered that the quality of Huggingface generation is slightly lower than vLLM, thus benchmarking using vLLM is recommended.\nWe are investigating the cause now.", "### 局限性 Limitations\n\n\n* 受限于模型规模,模型可能出现幻觉性问题。其中由于DPO模型生成的回复内容更长,更容易出现幻觉。我们也将持续进行MiniCPM模型的迭代改进;\n* 为了保证在学术研究用途上模型的通用性,我们未对模型进行任何身份认同训练。同时由于我们用ShareGPT开源语料作为部分训练数据,模型可能会输出类似GPT系列模型的身份认同信息;\n* 受限于模型规模,模型的输出受到提示词(prompt)的影响较大,可能多次尝试产生不一致的结果;\n* 受限于模型容量,模型的知识记忆较不准确,后续我们将结合RAG方法来增强模型的知识记忆能力。\n* Due to limitations in model size, the model may experience hallucinatory issues. As DPO model tend to generate longer response, hallucinations are more likely to occur. We will also continue to iterate and improve the MiniCPM model.\n* To ensure the universality of the model for academic research purposes, we did not conduct any identity training on the model. Meanwhile, as we use ShareGPT open-source corpus as part of the training data, the model may output identity information similar to the GPT series models.\n* Due to the limitation of model size, the output of the model is greatly influenced by prompt words, which may result in inconsistent results from multiple attempts.\n* Due to limited model capacity, the model's knowledge memory is not accurate. In the future, we will combine the RAG method to enhance the model's knowledge memory ability.\n\n\n模型下载 Download\n-------------\n\n\nHuggingFace: sft-bf16, ModelScope: sft-bf16, WiseModel: sft-bf16\nHuggingFace: sft-fp32, ModelScope: sft-fp32, WiseModel: sft-fp32\nHuggingFace: dpo-bf16, ModelScope: dpo-bf16, WiseModel: dpo-bf16\nHuggingFace: dpo-fp16, ModelScope: dpo-fp16, WiseModel: dpo-fp16\nHuggingFace: dpo-fp32, ModelScope: dpo-fp32, WiseModel: dpo-fp32\n\n\n模型使用 Usage\n----------\n\n\n* 安装'transformers>=4.36.0'以及'accelerate'后,运行以下代码\n* 注意:需要在'from\\_pretrained'中明确指明模型的数据类型,否则会引起较大计算误差\n* Run the following code after install 'transformers>=4.36.0' and 'accelerate'\n* Warning: It is necessary to specify the data type of the model clearly in 'from\\_pretrained', otherwise large calculation errors will be caused\n* 期望输出 Expected Output\n\n\n开源协议 LICENSE\n------------", "#### 模型协议 Model LICENSE\n\n\n* 本仓库中代码依照 Apache-2.0 协议开源\n* MiniCPM 模型权重的使用则需要遵循 “通用模型许可协议-来源说明-宣传限制-商业授权”。\n* MiniCPM 模型权重对学术研究完全开放。\n* 如需将模型用于商业用途,请联系[email protected]来获取书面授权,在登记后亦允许免费商业使用。\n* This repository is released under the Apache-2.0 License.\n* The usage of MiniCPM model weights must strictly follow the General Model License (GML).\n* The models and weights of MiniCPM are completely free for academic research.\n* If you intend to utilize the model for commercial purposes, please reach out to cpm@URL to obtain the certificate of authorization.", "#### 声明 Statement\n\n\n* 作为一个语言模型,MiniCPM 通过学习大量的文本来生成内容,但它无法理解、表达个人观点或价值判断,它所输出的任何内容都不代表模型开发者的观点和立场。\n* 因此用户在使用 MiniCPM 生成的内容时,应自行负责对其进行评估和验证。\n* 如果由于使用 MinCPM 开源模型而导致的任何问题,包括但不限于数据安全问题、公共舆论风险,或模型被误导、滥用、传播或不当利用所带来的任何风险和问题,我们将不承担任何责任。\n* As a language model, MiniCPM generates content by learning from a vast amount of text.\n* However, it does not possess the ability to comprehend or express personal opinions or value judgments.\n* Any content generated by MiniCPM does not represent the viewpoints or positions of the model developers.\n* Therefore, when using content generated by MiniCPM, users should take full responsibility for evaluating and verifying it on their own.\n\n\n\n工作引用 Citation\n-------------\n\n\n* 如果觉得MiniCPM有助于您的工作,请考虑引用下列技术报告\n* Please cite our techinical report if you find our work valuable." ]
[ "TAGS\n#transformers #MiniCPM #ModelBest #THUNLP #en #zh #endpoints_compatible #region-us \n", "### 评测结果 Evaluation Results\n\n\n详细的评测结果位于github仓库\n\n\nDetailed evaluation results are in github repo\n\n\n注意:我们发现使用Huggingface生成质量略差于vLLM,因此推荐使用vLLM进行测试。我们正在排查原因。\n\n\nNotice: We discovered that the quality of Huggingface generation is slightly lower than vLLM, thus benchmarking using vLLM is recommended.\nWe are investigating the cause now.", "### 局限性 Limitations\n\n\n* 受限于模型规模,模型可能出现幻觉性问题。其中由于DPO模型生成的回复内容更长,更容易出现幻觉。我们也将持续进行MiniCPM模型的迭代改进;\n* 为了保证在学术研究用途上模型的通用性,我们未对模型进行任何身份认同训练。同时由于我们用ShareGPT开源语料作为部分训练数据,模型可能会输出类似GPT系列模型的身份认同信息;\n* 受限于模型规模,模型的输出受到提示词(prompt)的影响较大,可能多次尝试产生不一致的结果;\n* 受限于模型容量,模型的知识记忆较不准确,后续我们将结合RAG方法来增强模型的知识记忆能力。\n* Due to limitations in model size, the model may experience hallucinatory issues. As DPO model tend to generate longer response, hallucinations are more likely to occur. We will also continue to iterate and improve the MiniCPM model.\n* To ensure the universality of the model for academic research purposes, we did not conduct any identity training on the model. Meanwhile, as we use ShareGPT open-source corpus as part of the training data, the model may output identity information similar to the GPT series models.\n* Due to the limitation of model size, the output of the model is greatly influenced by prompt words, which may result in inconsistent results from multiple attempts.\n* Due to limited model capacity, the model's knowledge memory is not accurate. In the future, we will combine the RAG method to enhance the model's knowledge memory ability.\n\n\n模型下载 Download\n-------------\n\n\nHuggingFace: sft-bf16, ModelScope: sft-bf16, WiseModel: sft-bf16\nHuggingFace: sft-fp32, ModelScope: sft-fp32, WiseModel: sft-fp32\nHuggingFace: dpo-bf16, ModelScope: dpo-bf16, WiseModel: dpo-bf16\nHuggingFace: dpo-fp16, ModelScope: dpo-fp16, WiseModel: dpo-fp16\nHuggingFace: dpo-fp32, ModelScope: dpo-fp32, WiseModel: dpo-fp32\n\n\n模型使用 Usage\n----------\n\n\n* 安装'transformers>=4.36.0'以及'accelerate'后,运行以下代码\n* 注意:需要在'from\\_pretrained'中明确指明模型的数据类型,否则会引起较大计算误差\n* Run the following code after install 'transformers>=4.36.0' and 'accelerate'\n* Warning: It is necessary to specify the data type of the model clearly in 'from\\_pretrained', otherwise large calculation errors will be caused\n* 期望输出 Expected Output\n\n\n开源协议 LICENSE\n------------", "#### 模型协议 Model LICENSE\n\n\n* 本仓库中代码依照 Apache-2.0 协议开源\n* MiniCPM 模型权重的使用则需要遵循 “通用模型许可协议-来源说明-宣传限制-商业授权”。\n* MiniCPM 模型权重对学术研究完全开放。\n* 如需将模型用于商业用途,请联系[email protected]来获取书面授权,在登记后亦允许免费商业使用。\n* This repository is released under the Apache-2.0 License.\n* The usage of MiniCPM model weights must strictly follow the General Model License (GML).\n* The models and weights of MiniCPM are completely free for academic research.\n* If you intend to utilize the model for commercial purposes, please reach out to cpm@URL to obtain the certificate of authorization.", "#### 声明 Statement\n\n\n* 作为一个语言模型,MiniCPM 通过学习大量的文本来生成内容,但它无法理解、表达个人观点或价值判断,它所输出的任何内容都不代表模型开发者的观点和立场。\n* 因此用户在使用 MiniCPM 生成的内容时,应自行负责对其进行评估和验证。\n* 如果由于使用 MinCPM 开源模型而导致的任何问题,包括但不限于数据安全问题、公共舆论风险,或模型被误导、滥用、传播或不当利用所带来的任何风险和问题,我们将不承担任何责任。\n* As a language model, MiniCPM generates content by learning from a vast amount of text.\n* However, it does not possess the ability to comprehend or express personal opinions or value judgments.\n* Any content generated by MiniCPM does not represent the viewpoints or positions of the model developers.\n* Therefore, when using content generated by MiniCPM, users should take full responsibility for evaluating and verifying it on their own.\n\n\n\n工作引用 Citation\n-------------\n\n\n* 如果觉得MiniCPM有助于您的工作,请考虑引用下列技术报告\n* Please cite our techinical report if you find our work valuable." ]
[ 32, 101, 655, 180, 262 ]
[ "passage: TAGS\n#transformers #MiniCPM #ModelBest #THUNLP #en #zh #endpoints_compatible #region-us \n### 评测结果 Evaluation Results\n\n\n详细的评测结果位于github仓库\n\n\nDetailed evaluation results are in github repo\n\n\n注意:我们发现使用Huggingface生成质量略差于vLLM,因此推荐使用vLLM进行测试。我们正在排查原因。\n\n\nNotice: We discovered that the quality of Huggingface generation is slightly lower than vLLM, thus benchmarking using vLLM is recommended.\nWe are investigating the cause now." ]
[ -0.007029905449599028, -0.18129052221775055, -0.005951940082013607, 0.035357579588890076, 0.09986112266778946, -0.0005699562607333064, 0.046559251844882965, 0.0894036665558815, -0.0011693296255543828, 0.014035903848707676, 0.1396910399198532, 0.05511273443698883, 0.004101620987057686, -0.008708887733519077, 0.04160221666097641, -0.15708091855049133, 0.06503809988498688, 0.045605048537254333, -0.060293711721897125, 0.0694865956902504, 0.167556032538414, -0.0876041129231453, 0.13962098956108093, 0.05953854322433472, -0.15145139396190643, 0.015299891121685505, 0.05207466706633568, 0.040434446185827255, 0.08226493000984192, 0.0536847747862339, -0.020794576033949852, 0.035934820771217346, -0.0060879685916006565, -0.08309119194746017, 0.047387413680553436, 0.07378146797418594, 0.03385411947965622, 0.00824330747127533, -0.0057003977708518505, 0.05097620189189911, 0.18805724382400513, -0.04968774691224098, -0.05933471396565437, 0.06705017387866974, -0.02407228946685791, -0.08207760751247406, -0.0734010636806488, 0.014700877480208874, 0.005885763559490442, 0.01351017877459526, 0.023399267345666885, 0.2902761697769165, -0.14518877863883972, 0.07959070056676865, 0.23669782280921936, -0.17718659341335297, -0.07245728373527527, 0.2521205246448517, 0.0753178521990776, -0.08139427751302719, -0.033760204911231995, 0.13844187557697296, 0.09468132257461548, -0.0026721120811998844, 0.0026779472827911377, 0.0028682169504463673, 0.019566699862480164, 0.036671992391347885, -0.0707022175192833, -0.04395657032728195, 0.21017861366271973, 0.024222983047366142, -0.01676894538104534, -0.07778884470462799, -0.04407808557152748, -0.03170435130596161, 0.019980154931545258, -0.056891828775405884, -0.0009387274621985853, 0.028179902583360672, -0.023554271087050438, -0.056964561343193054, -0.05591902509331703, -0.06277237087488174, -0.050856560468673706, 0.07139258831739426, 0.012281373143196106, 0.06602782756090164, -0.12072503566741943, 0.07584036141633987, -0.016529422253370285, -0.06980713456869125, -0.06994563341140747, -0.15332739055156708, 0.09148465096950531, 0.0343741737306118, -0.029552128165960312, -0.020706990733742714, 0.15407049655914307, 0.06227974221110344, 0.09200252592563629, -0.009465874172747135, 0.01687920093536377, -0.006674732081592083, 0.030380485579371452, 0.16072417795658112, -0.0959244966506958, -0.03150952234864235, 0.11796808242797852, -0.006341414991766214, 0.06780342012643814, -0.010086788795888424, 0.02585897408425808, -0.10685950517654419, 0.0907537117600441, 0.09983225166797638, -0.005054982379078865, 0.08520478010177612, -0.06697309762239456, -0.07043292373418808, -0.13299769163131714, -0.048997700214385986, -0.02080552838742733, -0.02179517038166523, -0.021712658926844597, 0.16008159518241882, 0.037784550338983536, -0.007054767571389675, -0.06459946185350418, -0.04314681142568588, -0.07845790684223175, -0.07143807411193848, -0.07053966075181961, -0.06252117455005646, 0.03135087341070175, -0.12207255512475967, 0.05156022682785988, -0.07779049873352051, -0.11634596437215805, -0.02535473369061947, 0.0038164493162184954, 0.009199460968375206, -0.039813365787267685, 0.056789007037878036, -0.047064024955034256, -0.01596369780600071, 0.005027248989790678, 0.08947260677814484, -0.038541000336408615, 0.02651931159198284, 0.042357850819826126, 0.031899482011795044, -0.08141711354255676, -0.019544366747140884, -0.028555387631058693, 0.04780369997024536, 0.06622783094644547, -0.036718253046274185, -0.10905038565397263, 0.1576538234949112, -0.07052676379680634, -0.06317166239023209, -0.043943632394075394, -0.0029794408474117517, 0.03134134039282799, 0.1293453723192215, -0.13065388798713684, -0.06242143735289574, 0.05081261321902275, -0.12080085277557373, -0.1050158143043518, 0.14760424196720123, 0.009458748623728752, 0.022331831976771355, 0.12191835045814514, -0.0022843179758638144, 0.125326007604599, -0.19225206971168518, -0.020844638347625732, 0.03509785607457161, -0.07853440195322037, 0.0455203503370285, 0.043436478823423386, 0.08068293333053589, -0.14074291288852692, 0.10252724587917328, -0.06006870046257973, 0.007547332439571619, -0.09189192205667496, -0.03843272849917412, -0.15482521057128906, -0.025032080709934235, -0.05366566404700279, 0.04065868258476257, 0.009939560666680336, -0.050259221345186234, -0.04092514142394066, -0.1113775223493576, 0.08663772791624069, -0.023160140961408615, -0.006348270457237959, -0.15282754600048065, 0.1418062448501587, -0.0006764328572899103, 0.03414156287908554, -0.009384505450725555, -0.0311939287930727, 0.01573129929602146, 0.04729769378900528, -0.0020969773177057505, 0.19795012474060059, 0.025106357410550117, -0.06771818548440933, 0.022668274119496346, 0.05191590636968613, -0.025603611022233963, -0.04619232565164566, 0.029351724311709404, -0.1542227417230606, -0.0030155712738633156, -0.04354240745306015, -0.10526778548955917, -0.013441141694784164, -0.02526128478348255, 0.12803754210472107, 0.18103109300136566, 0.03463297337293625, 0.009342518635094166, 0.035754818469285965, -0.04776196554303169, -0.03296038880944252, -0.06930872052907944, 0.05342394486069679, -0.035693034529685974, -0.040911924093961716, 0.0916159525513649, 0.06284727156162262, 0.16049063205718994, 0.12458932399749756, -0.04375477880239487, 0.10487950593233109, 0.03939517214894295, -0.054152533411979675, 0.055077169090509415, -0.022789044305682182, -0.0593547560274601, 0.07723882794380188, 0.0013175575295463204, 0.043476928025484085, -0.0795559361577034, -0.021214991807937622, 0.011689210310578346, -0.05022359639406204, -0.049612510949373245, 0.07128696888685226, 0.06985030323266983, 0.018150771036744118, -0.0025405569467693567, 0.052537064999341965, 0.03218483179807663, 0.21697549521923065, -0.07065971195697784, 0.01958327926695347, -0.018351927399635315, -0.03190876170992851, -0.029256965965032578, 0.06542211771011353, -0.20291361212730408, -0.05130786448717117, 0.04685044288635254, -0.034627415239810944, 0.08495761454105377, -0.09145545959472656, -0.00968084391206503, 0.010166168212890625, -0.01712985895574093, 0.04534405469894409, 0.1705799400806427, -0.018720511347055435, 0.06875555217266083, -0.05606083944439888, 0.10785147547721863, -0.06280707567930222, -0.029891222715377808, -0.05139126256108284, 0.09397685527801514, -0.052178725600242615, -0.1539200097322464, -0.09655670821666718, 0.03096357360482216, -0.15777359902858734, -0.024116067215800285, -0.01893104426562786, -0.20700260996818542, -0.09705216437578201, -0.0003875589754898101, 0.28305497765541077, -0.03400006517767906, 0.06575983017683029, 0.1628442257642746, 0.06021367385983467, 0.014382950030267239, -0.19300132989883423, -0.06491397321224213, -0.01619957946240902, -0.10169330984354019, 0.03864741697907448, -0.06528447568416595, 0.04829725623130798, 0.04979325458407402, 0.030524758622050285, 0.05394037067890167, -0.007211975287646055, 0.32132506370544434, -0.01992679201066494, -0.03821587562561035, 0.36082324385643005, 0.10184833407402039, -0.039524469524621964, 0.05349263548851013, -0.0064017074182629585, -0.14367203414440155, 0.00786982849240303, 0.03457683324813843, -0.07522208243608475, -0.10467099398374557, -0.08188129216432571, -0.08901073783636093, -0.004968418274074793, -0.018996141850948334, 0.03418242558836937, -0.08056186139583588, 0.04029500111937523, -0.03882334381341934, 0.09780184924602509, -0.02714279480278492, 0.03977955877780914, 0.1137450635433197, -0.0636458545923233, 0.09473677724599838, -0.043208952993154526, -0.028461836278438568, 0.11854533851146698, -0.0039004001300781965, 0.1605968028306961, -0.05560247227549553, -0.01632627286016941, 0.1513911783695221, 0.07536110281944275, 0.052956126630306244, 0.04650551453232765, -0.09035848081111908, -0.05243884399533272, -0.016077518463134766, -0.02378683164715767, -0.1326739490032196, -0.03452671691775322, 0.010660284198820591, -0.03295145183801651, -0.07800742983818054, 0.1525101512670517, 0.07630597054958344, 0.1832588016986847, 0.04941178485751152, -0.1264776587486267, -0.055462438613176346, 0.05326181277632713, -0.008437084965407848, -0.051986631006002426, -0.0027289516292512417, 0.021061580628156662, -0.06870332360267639, 0.07071825861930847, -0.03226818889379501, 0.09059399366378784, -0.04277632758021355, 0.08795687556266785, -0.10145540535449982, 0.05973893031477928, 0.029581494629383087, 0.06363088637590408, -0.19256146252155304, 0.11430851370096207, 0.010978278703987598, 0.09448982030153275, -0.08606318384408951, 0.0075591071508824825, 0.07585456222295761, 0.15848422050476074, 0.07580620050430298, 0.018830114975571632, 0.028623787686228752, 0.030678128823637962, -0.046929821372032166, 0.10647550970315933, 0.02444637008011341, 0.018736455589532852, 0.048744652420282364, -0.13624097406864166, 0.06992737203836441, -0.005191147793084383, 0.1014765128493309, -0.19799570739269257, 0.0066639515571296215, 0.02145177498459816, -0.006088297814130783, 0.019627736881375313, -0.10134831815958023, -0.01989709958434105, 0.02259662002325058, 0.1600578874349594, -0.031674597412347794, -0.02505546063184738, -0.10328514128923416, 0.13805516064167023, 0.07889293134212494, -0.03296839818358421, -0.037489306181669235, -0.09881174564361572, 0.14158830046653748, 0.016290584579110146, -0.16540241241455078, -0.08646079897880554, -0.10454490035772324, -0.11585325747728348, 0.027800647541880608, 0.12482665479183197, -0.13428716361522675, 0.012853509746491909, 0.0667235404253006, -0.04668804630637169, -0.07104019820690155, -0.13363845646381378, 0.03306398540735245, 0.18510258197784424, -0.17520250380039215, 0.03133249282836914, -0.01204567402601242, -0.03713098168373108, 0.030830876901745796, -0.031123245134949684, 0.060273922979831696, 0.23087160289287567, -0.014127736911177635, 0.08515125513076782, -0.0011738878674805164, -0.02978982776403427, -0.2783195376396179, -0.042848993092775345, 0.013759949244558811, 0.07922997325658798, -0.09668739885091782, 0.002893011551350355, 0.1082424744963646, -0.057692717760801315, 0.017866728827357292, -0.023136377334594727, -0.23526626825332642, -0.11276296526193619, 0.02768295630812645, 0.04173368960618973, 0.3718765676021576, -0.005618828348815441, -0.018679693341255188, -0.08236056566238403, -0.2331053763628006, 0.09754981845617294, -0.09003683924674988, 0.07901791483163834, -0.09090468287467957, -0.006996639538556337, 0.0546121746301651, -0.04311290755867958, 0.16560278832912445, -0.13848967850208282, 0.09249752014875412, -0.05166751146316528, -0.046906907111406326, 0.07662864029407501, -0.04660817235708237, 0.08431723713874817, -0.00004227057070238516, 0.07783340662717819, -0.10475700348615646, -0.039542991667985916, -0.10952022671699524, 0.11239190399646759, -0.010712068527936935, -0.06655392795801163, -0.14847688376903534, 0.054326292127370834, -0.013305011205375195, 0.007040245458483696, 0.10857025533914566, -0.026810117065906525, 0.006204573903232813, 0.07563278824090958, 0.09349015355110168, -0.25899824500083923, 0.034354791045188904, 0.03968802094459534, 0.04425780847668648, 0.10542070120573044, -0.16066081821918488, 0.07335808128118515, 0.08233226835727692, 0.005068870726972818, -0.00943797454237938, 0.0298149436712265, -0.04710002616047859, 0.006641503889113665, 0.13628563284873962, -0.15587253868579865, -0.04959556460380554, -0.047705113887786865, -0.060472521930933, 0.05329377204179764, 0.046486590057611465, 0.22370930016040802, -0.018752489238977432, 0.011128068901598454, -0.0049996827729046345, 0.003121020970866084, -0.05447318032383919, 0.12865500152111053, 0.03850371018052101, 0.02667682059109211, -0.13138838112354279, 0.04617862030863762, -0.08798189461231232, 0.07230368256568909, -0.012639796361327171, -0.07910933345556259, -0.08561225980520248, -0.04239198938012123, 0.009966548532247543, 0.09304729849100113, -0.03591559827327728, -0.029956577345728874, -0.13225248456001282, -0.1374678760766983, 0.0008156708790920675, 0.003170343115925789, 0.08875831961631775, 0.11562928557395935, -0.054417435079813004, -0.04701881855726242, -0.0401037223637104, 0.004986949265003204, 0.04682082682847977, -0.017912017181515694, -0.21080687642097473, 0.08802185952663422, -0.004020586609840393, 0.10052767395973206, -0.09412624686956406, -0.002272270852699876, -0.1231643483042717, -0.020640898495912552, -0.18347449600696564, -0.04763184115290642, 0.026195168495178223, -0.01202356070280075, 0.0669357180595398, -0.0617840401828289, -0.15457159280776978, 0.0720461755990982, -0.07213762402534485, 0.013440023176372051, -0.020221391692757607, 0.05134095251560211, 0.011415361426770687, 0.029058363288640976, 0.1242644339799881, -0.04318128526210785, 0.07616350054740906, -0.030477555468678474, -0.021776655688881874, 0.045308686792850494, -0.18348194658756256, -0.05518330633640289, 0.08187967538833618, 0.04886586219072342, -0.019504666328430176, 0.023459650576114655, 0.07082602381706238, 0.024681204929947853, 0.00820586085319519, 0.030938245356082916, 0.06416723132133484, -0.05277709290385246, -0.0855041965842247, 0.007049820851534605, 0.008752464316785336, 0.020142750814557076, -0.022155463695526123, 0.04487167298793793, 0.12529303133487701, 0.09726529568433762, 0.014684763737022877, -0.0194696094840765, -0.16838698089122772, 0.00830417312681675, -0.00984482653439045, -0.10293831676244736, -0.028975795954465866, 0.0021882206201553345, 0.06077804043889046, -0.014182863757014275, 0.2449350357055664, 0.11702872812747955, 0.005042670294642448, 0.007772746030241251, 0.18846075236797333, -0.029718095436692238, 0.0003312989429105073, 0.11638941615819931, 0.004579353611916304, 0.13689325749874115, -0.053380146622657776, -0.034658756107091904, 0.06998783349990845, -0.09110356122255325, 0.14839224517345428, 0.04822993278503418, -0.035704005509614944, 0.08797920495271683, 0.05165775865316391, 0.008065287955105305, 0.022829916328191757, 0.014720307663083076, -0.04585376754403114, 0.12946702539920807, -0.03044453077018261, -0.12339377403259277, 0.13701024651527405, -0.09897657483816147, 0.01810203492641449, -0.02869013510644436, -0.006283747963607311, -0.1463201940059662, -0.12284358590841293, -0.07512173801660538, -0.20523324608802795, 0.10039401054382324, -0.10859007388353348, -0.016365915536880493, -0.08122794330120087, 0.05900624021887779, -0.07911669462919235, 0.08075745403766632, -0.08636533468961716, -0.11989813297986984, 0.0785241648554802, -0.024483587592840195, -0.10325786471366882, -0.011768877506256104, 0.03858496621251106, -0.04452567920088768, 0.027312010526657104, -0.02097550593316555, 0.025646530091762543, -0.02509469911456108, -0.08288910984992981, -0.02417762018740177, -0.027694258838891983, -0.06983157992362976, -0.05789756029844284, -0.088569775223732, 0.10396303236484528, -0.028263943269848824, -0.016770722344517708, -0.014409912750124931, 0.16765199601650238, 0.022424353286623955, -0.06278514117002487, -0.12636616826057434, 0.007873471826314926, -0.054353103041648865, 0.03975142538547516, -0.009976043365895748, 0.028603317216038704, -0.04908722639083862, 0.29296132922172546, 0.11499285697937012, -0.06207657605409622, 0.00033675841405056417, -0.07435646653175354, 0.020152198150753975, -0.06945610791444778, 0.17368726432323456, 0.020711664110422134, 0.12320763617753983, -0.03448185324668884, 0.04327443242073059, -0.04210294038057327, -0.029141951352357864, -0.062144361436367035, 0.07125066965818405, 0.07137252390384674, -0.046930860728025436, -0.027866339311003685, 0.03252615034580231, -0.046302780508995056, -0.0053444840013980865, 0.06731386482715607, -0.15821880102157593, -0.046171799302101135, -0.03454064205288887, 0.049210429191589355, -0.06873206794261932, 0.09366195648908615, -0.08513263612985611, 0.03372931480407715, 0.02272685244679451, -0.019690684974193573, -0.1433858722448349, -0.0600271075963974, 0.12623290717601776, 0.1163797527551651, 0.1186923161149025, -0.032800983637571335, -0.03777356073260307, 0.07561182230710983, -0.10685569792985916, -0.08995477855205536, -0.019915208220481873, -0.03216233849525452, -0.12198159843683243, -0.039367012679576874, 0.07561811804771423, 0.02740832418203354, 0.0023442672099918127, 0.02984943799674511, 0.011059924028813839, -0.014475255273282528, 0.12300413101911545, 0.040855977684259415, -0.03510741516947746, 0.13164114952087402, -0.14435340464115143, 0.08194208890199661, 0.1004456952214241, -0.04326357692480087, -0.026686301454901695, -0.11590644717216492, 0.17826007306575775, -0.006448836997151375, 0.01886652782559395, 0.00009278854849981144, -0.1277666836977005, -0.003054513595998287, 0.15814904868602753, -0.07905298471450806, -0.019775869324803352, -0.12718722224235535, -0.02167162485420704, -0.001297742361202836, -0.026628093793988228, 0.13205231726169586, 0.07088743150234222, 0.03847820311784744, 0.053234636783599854, -0.15350842475891113, 0.053478408604860306, -0.016429493203759193, -0.06174127385020256, -0.08909101039171219 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text-generation
aidonuts/marigold-006-ep4
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T13:28:47+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 60, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.04654794931411743, 0.16618601977825165, -0.005445904564112425, 0.01853804849088192, 0.0981811136007309, 0.011998992413282394, 0.06433123350143433, 0.11398410052061081, -0.0230073444545269, 0.11406639218330383, 0.03047988750040531, 0.10172267258167267, 0.11317981779575348, 0.14841650426387787, -0.002152352826669812, -0.22403094172477722, 0.050844956189394, -0.12105348706245422, -0.033293843269348145, 0.11749980598688126, 0.1483822613954544, -0.09928343445062637, 0.07274559140205383, -0.029687678441405296, -0.012143402360379696, -0.030057786032557487, -0.05890674889087677, -0.046214159578084946, 0.04651786759495735, 0.06640566885471344, 0.06770290434360504, 0.0071083661168813705, 0.09012923389673233, -0.2696533799171448, 0.018959321081638336, 0.07145345956087112, -0.002759667346253991, 0.06957992166280746, 0.06404146552085876, -0.07107418030500412, 0.10337356477975845, -0.05106033384799957, 0.14650006592273712, 0.08365883678197861, -0.09081148356199265, -0.1895141303539276, -0.08866965025663376, 0.09882009029388428, 0.17572562396526337, 0.04925641790032387, -0.02320658043026924, 0.09761467576026917, -0.08769196271896362, 0.015438909642398357, 0.04981724172830582, -0.07620415836572647, -0.05378096550703049, 0.05986575037240982, 0.07907199114561081, 0.06627275794744492, -0.12434766441583633, -0.02885502204298973, 0.005009706597775221, 0.010980482213199139, 0.0769270583987236, 0.01728810742497444, 0.146672785282135, 0.0338633768260479, -0.12615777552127838, -0.04880760237574577, 0.09869225323200226, 0.03395522013306618, -0.04422314465045929, -0.24749068915843964, -0.03152675926685333, -0.030810698866844177, -0.029386121779680252, -0.03716538846492767, 0.04340358078479767, -0.007673026993870735, 0.08638741075992584, -0.0060646249912679195, -0.07403432577848434, -0.03937075287103653, 0.06169692054390907, 0.0672287791967392, 0.02999979443848133, -0.013745363801717758, 0.010938193649053574, 0.11620724946260452, 0.1095694974064827, -0.12054188549518585, -0.05555335059762001, -0.06393084675073624, -0.08656639605760574, -0.040790557861328125, 0.034162238240242004, 0.03456587344408035, 0.05349370837211609, 0.25305667519569397, 0.015654386952519417, 0.059652652591466904, 0.034477248787879944, 0.007892133668065071, 0.05848940089344978, 0.11044429242610931, -0.06018859148025513, -0.10444226115942001, -0.02648012898862362, 0.08843598514795303, 0.008199662901461124, -0.03287925571203232, -0.05088530853390694, 0.06019928678870201, 0.01946467161178589, 0.11926145106554031, 0.09061790257692337, 0.010536285117268562, -0.07121123373508453, -0.061038948595523834, 0.1891259253025055, -0.16544590890407562, 0.04322727024555206, 0.035097137093544006, -0.03903156518936157, 0.00019933005387429148, 0.013914269395172596, 0.016625655815005302, -0.025983380153775215, 0.09017423540353775, -0.054113563150167465, -0.04145489260554314, -0.11186197400093079, -0.03383193537592888, 0.033762916922569275, 0.008953776210546494, -0.035059962421655655, -0.033713940531015396, -0.08351044356822968, -0.07577689737081528, 0.09320491552352905, -0.07346344739198685, -0.04878907650709152, -0.01804324984550476, -0.07530532777309418, 0.022395428270101547, 0.019394835457205772, 0.07707412540912628, -0.02362251654267311, 0.04399976506829262, -0.05189276114106178, 0.05863580107688904, 0.11207318305969238, 0.03570080175995827, -0.05736649036407471, 0.06062258034944534, -0.23834340274333954, 0.09552820026874542, -0.07409077137708664, 0.05591456592082977, -0.153293639421463, -0.024439791217446327, 0.04788333550095558, 0.008784620091319084, -0.009650949388742447, 0.13416339457035065, -0.21702027320861816, -0.02536402828991413, 0.1717337965965271, -0.10057014971971512, -0.07069246470928192, 0.05619903281331062, -0.04835370555520058, 0.10988964140415192, 0.03825836628675461, -0.025690359994769096, 0.06171267107129097, -0.1267417073249817, 0.003717758459970355, -0.05005312338471413, -0.017048977315425873, 0.1548657864332199, 0.07182947546243668, -0.07217690348625183, 0.07399354875087738, 0.025708531960844994, -0.0246540866792202, -0.04625825211405754, -0.015164627693593502, -0.10536660254001617, 0.014689887873828411, -0.06369215250015259, 0.014470234513282776, -0.020807426422834396, -0.09071163833141327, -0.027962757274508476, -0.17504668235778809, -0.03014434315264225, 0.08651752024888992, -0.008693269453942776, -0.01803150773048401, -0.1178668737411499, 0.009341353550553322, 0.04177580401301384, 0.0061247628182172775, -0.13462838530540466, -0.04812471568584442, 0.02780051715672016, -0.1600649207830429, 0.034652888774871826, -0.05392369255423546, 0.04932025074958801, 0.025790516287088394, -0.028889117762446404, -0.026493212208151817, 0.021633783355355263, 0.005992184858769178, -0.011999987065792084, -0.24343903362751007, -0.028118690475821495, -0.024888472631573677, 0.1682123839855194, -0.20917098224163055, 0.03546025976538658, 0.07867541164159775, 0.15366052091121674, 0.011240328662097454, -0.04177491366863251, 0.005974748637527227, -0.06935794651508331, -0.02736494317650795, -0.05875484645366669, -0.0047869328409433365, -0.03310677409172058, -0.04545191675424576, 0.04568447172641754, -0.16510973870754242, -0.032636504620313644, 0.09776268899440765, 0.06289951503276825, -0.13922683894634247, -0.020621931180357933, -0.03630133345723152, -0.049253206700086594, -0.04911839962005615, -0.0605199858546257, 0.10893940925598145, 0.05891856551170349, 0.04574795812368393, -0.05928509309887886, -0.07568105310201645, -0.001827909960411489, -0.013898161239922047, -0.017864689230918884, 0.09759635478258133, 0.0751434788107872, -0.13251115381717682, 0.09224759042263031, 0.09603385627269745, 0.07919023185968399, 0.09113933145999908, -0.02355697751045227, -0.08261934667825699, -0.045987509191036224, 0.031442027539014816, 0.020124373957514763, 0.13039541244506836, -0.024294709786772728, 0.04352088272571564, 0.042134687304496765, -0.019369594752788544, 0.014752166345715523, -0.08687400817871094, 0.033972494304180145, 0.028472330421209335, -0.016721390187740326, 0.050190530717372894, -0.03876714035868645, 0.02440318465232849, 0.08830609917640686, 0.045322712510824203, 0.03507532551884651, 0.015493292361497879, -0.05206458270549774, -0.1083620935678482, 0.16405931115150452, -0.12714070081710815, -0.22483378648757935, -0.13936103880405426, 0.0037376401014626026, 0.035628627985715866, -0.015835661441087723, 0.002417160663753748, -0.059374887496232986, -0.12220635265111923, -0.08858037739992142, 0.015140829607844353, 0.04942670464515686, -0.09028962254524231, -0.06437795609235764, 0.058117836713790894, 0.03889724239706993, -0.14560972154140472, 0.017612040042877197, 0.04854894429445267, -0.09789852797985077, -0.006774199660867453, 0.08094939589500427, 0.0698540136218071, 0.1770169734954834, 0.017703235149383545, -0.021850809454917908, 0.032354529947042465, 0.20614571869373322, -0.13538233935832977, 0.11083246022462845, 0.13607586920261383, -0.09041404724121094, 0.08072979003190994, 0.19951270520687103, 0.03932560607790947, -0.10153959691524506, 0.031980328261852264, 0.02283124253153801, -0.0284719280898571, -0.24526868760585785, -0.07212468236684799, -0.004402178805321455, -0.058010730892419815, 0.07660572230815887, 0.09286724030971527, 0.08215958625078201, 0.012304253876209259, -0.09310996532440186, -0.08154371380805969, 0.05942574888467789, 0.10367169976234436, 0.024584239348769188, -0.010839897207915783, 0.08998730033636093, -0.034100502729415894, 0.019626356661319733, 0.0853661298751831, 0.005239574704319239, 0.17840281128883362, 0.05159219726920128, 0.18830420076847076, 0.07925192266702652, 0.07219027727842331, 0.009912233799695969, 0.013080619275569916, 0.018877580761909485, 0.03300119563937187, -0.002769160782918334, -0.08440786600112915, -0.02248465269804001, 0.11566436290740967, 0.06668911874294281, 0.010815348476171494, 0.015172341838479042, -0.04104290530085564, 0.07965951412916183, 0.1831512451171875, -0.007656289264559746, -0.1783534437417984, -0.057547420263290405, 0.07553383708000183, -0.09879875183105469, -0.09854305535554886, -0.013454320840537548, 0.03072015568614006, -0.17046253383159637, 0.023390959948301315, -0.02239842526614666, 0.1106182336807251, -0.14194999635219574, -0.020490378141403198, 0.07218493521213531, 0.07199500501155853, 0.004729843698441982, 0.05758659541606903, -0.16417601704597473, 0.10671813786029816, 0.008950476534664631, 0.06779605895280838, -0.09610627591609955, 0.1008887067437172, -0.004196076653897762, -0.02063460275530815, 0.1393408179283142, 0.002700034761801362, -0.06884108483791351, -0.0763031542301178, -0.08754398673772812, -0.009632662869989872, 0.12754282355308533, -0.1419651061296463, 0.08767123520374298, -0.037212442606687546, -0.0424150750041008, -0.0017086371080949903, -0.10206665843725204, -0.11638247221708298, -0.18888559937477112, 0.06001543253660202, -0.13492922484874725, 0.03152317553758621, -0.10799519717693329, -0.032371897250413895, -0.030304040759801865, 0.19337286055088043, -0.23447458446025848, -0.07199826091527939, -0.1475764364004135, -0.10233612358570099, 0.1443224400281906, -0.0501345656812191, 0.08485390990972519, -0.007241467013955116, 0.16846685111522675, 0.019060896709561348, -0.02531743235886097, 0.0971490666270256, -0.09173708409070969, -0.19302815198898315, -0.07869284600019455, 0.15662524104118347, 0.13260218501091003, 0.031680017709732056, -0.002461588243022561, 0.036563750356435776, -0.015421539545059204, -0.11935004591941833, 0.015969349071383476, 0.1787186712026596, 0.06237189099192619, 0.02331034652888775, -0.027346095070242882, -0.11273157596588135, -0.06900003552436829, -0.028530338779091835, 0.03054865077137947, 0.17762407660484314, -0.07057618349790573, 0.18207968771457672, 0.14163152873516083, -0.05922834202647209, -0.20400173962116241, 0.010538800619542599, 0.03055560030043125, 0.0009220078936778009, 0.02591954916715622, -0.20123432576656342, 0.08688826113939285, 0.004683020059019327, -0.05110127478837967, 0.13194532692432404, -0.17217805981636047, -0.14451217651367188, 0.0765485092997551, 0.038384392857551575, -0.19559739530086517, -0.12913893163204193, -0.09174312651157379, -0.045869920402765274, -0.18591414391994476, 0.09569250047206879, 0.0305706188082695, 0.010893458500504494, 0.03030681423842907, 0.029179483652114868, 0.019487828016281128, -0.0418255440890789, 0.18391458690166473, -0.024792250245809555, 0.026594700291752815, -0.08539514988660812, -0.06927408277988434, 0.03743394836783409, -0.052842434495687485, 0.07349982857704163, -0.023486759513616562, 0.007861839607357979, -0.10348054021596909, -0.042148489505052567, -0.03735732287168503, 0.015448716469109058, -0.09657872468233109, -0.08514349907636642, -0.045032672584056854, 0.09675803780555725, 0.09690850973129272, -0.033646680414676666, -0.028050623834133148, -0.07533035427331924, 0.04412057250738144, 0.19926515221595764, 0.1785389482975006, 0.042153384536504745, -0.08034496754407883, -0.004150947090238333, -0.010121207684278488, 0.04310847446322441, -0.20463712513446808, 0.06283636391162872, 0.05450061708688736, 0.01973269321024418, 0.11436162889003754, -0.019565396010875702, -0.15359151363372803, -0.07263088971376419, 0.06303015351295471, -0.060181066393852234, -0.19620554149150848, 0.00867035984992981, 0.060603946447372437, -0.16371412575244904, -0.04535605385899544, 0.04643881320953369, -0.005620351992547512, -0.038163937628269196, 0.021896906197071075, 0.09194854646921158, 0.0026654244866222143, 0.07427921891212463, 0.05387866869568825, 0.0827430784702301, -0.10537070035934448, 0.08090532571077347, 0.08839722722768784, -0.08452684432268143, 0.023530138656497, 0.10478579998016357, -0.059433579444885254, -0.03440561518073082, 0.020135708153247833, 0.08153781294822693, 0.01775863952934742, -0.040019966661930084, 0.013229827396571636, -0.10452935844659805, 0.05954122915863991, 0.08839859813451767, 0.032507482916116714, 0.016702456399798393, 0.03425082191824913, 0.04607953503727913, -0.07238735258579254, 0.12142276018857956, 0.031868141144514084, 0.017129309475421906, -0.036505792289972305, -0.040896978229284286, 0.019542274996638298, -0.03214648738503456, -0.005015232600271702, -0.03023446537554264, -0.07695909589529037, -0.014793801121413708, -0.1626158058643341, -0.011131818406283855, -0.05648450180888176, 0.010329355485737324, 0.03204665705561638, -0.032609567046165466, 0.008124498650431633, 0.009250079281628132, -0.07695289701223373, -0.0663459524512291, -0.020460480824112892, 0.09540658444166183, -0.16213038563728333, 0.022481130436062813, 0.08244425803422928, -0.12187694013118744, 0.09281346201896667, 0.016204802319407463, -0.006236857734620571, 0.025038830935955048, -0.1475188434123993, 0.034843120723962784, -0.03386561945080757, 0.010836300440132618, 0.04373383894562721, -0.21569781005382538, -0.00004886732858722098, -0.033673107624053955, -0.06639216095209122, -0.009451326914131641, -0.03672455996274948, -0.11508306115865707, 0.1058407872915268, 0.007236586883664131, -0.08753558248281479, -0.03186136856675148, 0.029325377196073532, 0.0838974118232727, -0.021959776058793068, 0.15145497024059296, -0.008370938710868359, 0.07429654151201248, -0.16209737956523895, -0.018623165786266327, -0.006028574425727129, 0.022658247500658035, -0.01664556935429573, -0.01111356820911169, 0.044031109660863876, -0.022746501490473747, 0.17925859987735748, -0.030318550765514374, 0.02272745408117771, 0.06815794110298157, 0.019072026014328003, -0.030184008181095123, 0.10406795144081116, 0.04094860330224037, 0.02014910988509655, 0.018591465428471565, 0.003289656015112996, -0.04647882282733917, -0.03173251822590828, -0.19407226145267487, 0.07288651913404465, 0.15608493983745575, 0.09729263186454773, -0.016707008704543114, 0.07954329252243042, -0.10199416428804398, -0.1109243705868721, 0.12477338314056396, -0.04797708988189697, -0.002418199321255088, -0.07150927931070328, 0.13247236609458923, 0.1437523066997528, -0.1859612911939621, 0.07269313186407089, -0.0699717253446579, -0.04708027467131615, -0.10980689525604248, -0.19441905617713928, -0.05561789125204086, -0.049456022679805756, -0.016053348779678345, -0.04698808491230011, 0.07504211366176605, 0.054538097232580185, 0.006766852922737598, -0.0023397188633680344, 0.06506035476922989, -0.031050674617290497, -0.0037882844917476177, 0.032597362995147705, 0.06591679900884628, 0.012734474614262581, -0.030802709981799126, 0.016619903966784477, -0.013545602560043335, 0.045626189559698105, 0.06578011065721512, 0.04976864159107208, -0.02938537672162056, 0.014603170566260815, -0.038539156317710876, -0.10249634087085724, 0.043612558394670486, -0.024421939626336098, -0.0789753645658493, 0.15477414429187775, 0.023680059239268303, 0.007779473438858986, -0.020137663930654526, 0.23901568353176117, -0.0738423764705658, -0.0964353010058403, -0.14737580716609955, 0.10557299107313156, -0.038081806153059006, 0.05800395458936691, 0.04625935107469559, -0.10226529091596603, 0.018044332042336464, 0.1338089406490326, 0.16182038187980652, -0.039008259773254395, 0.020095856860280037, 0.031135575845837593, 0.00566398398950696, -0.03622615709900856, 0.04847532883286476, 0.06906453520059586, 0.16569648683071136, -0.04632584750652313, 0.09100406616926193, 0.0019041687482967973, -0.09579581767320633, -0.038361791521310806, 0.11069868505001068, -0.016052277758717537, 0.019335128366947174, -0.05818064883351326, 0.11742528527975082, -0.06386786699295044, -0.23783175647258759, 0.06453443318605423, -0.0684293657541275, -0.13765870034694672, -0.02378307841718197, 0.08207765966653824, -0.012955902144312859, 0.027587108314037323, 0.0730307325720787, -0.07240920513868332, 0.201939657330513, 0.03798431158065796, -0.05499868467450142, -0.055047210305929184, 0.0805421993136406, -0.10008571296930313, 0.2739645540714264, 0.01557221356779337, 0.04601577669382095, 0.10384146869182587, -0.009341772645711899, -0.13838784396648407, 0.019836371764540672, 0.09581108391284943, -0.10502193123102188, 0.04196618124842644, 0.19815568625926971, -0.0014755994779989123, 0.12389086186885834, 0.07657600939273834, -0.07551808655261993, 0.0478031262755394, -0.08054235577583313, -0.06760486960411072, -0.09260394424200058, 0.09703279286623001, -0.07772123068571091, 0.14251399040222168, 0.13876807689666748, -0.05074559152126312, 0.012724342755973339, -0.031311117112636566, 0.044293127954006195, -0.00010600237874314189, 0.10321761667728424, 0.004272161517292261, -0.1832672357559204, 0.024692710489034653, 0.005650998093187809, 0.10749758034944534, -0.16033467650413513, -0.09566054493188858, 0.042343202978372574, 0.003505636239424348, -0.0672195628285408, 0.1290110945701599, 0.05665452033281326, 0.04342988133430481, -0.03997718170285225, -0.03521440550684929, -0.0060732318088412285, 0.13561366498470306, -0.10713256150484085, 0.0009933578548952937 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
null
jeevana/GenAI_QnA_Mistral7b_QLoRA_G8_FV04
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
2024-02-10T13:30:01+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 31, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.06646376848220825, 0.2168014943599701, -0.00225935154594481, 0.023818302899599075, 0.1271018385887146, -0.001635765191167593, 0.04218708351254463, 0.13324736058712006, -0.020175931975245476, 0.11144465953111649, 0.046588581055402756, 0.09377603232860565, 0.09928803145885468, 0.18404334783554077, 0.04859916493296623, -0.2059975117444992, 0.007056170143187046, -0.09090408682823181, 0.014076028019189835, 0.1116579994559288, 0.13719257712364197, -0.10291384905576706, 0.08272874355316162, -0.04045208916068077, -0.02019004337489605, 0.00012576708104461432, -0.09259183704853058, -0.07032395154237747, 0.06885425746440887, 0.06264153122901917, 0.051234472543001175, 0.001456156256608665, 0.09140396863222122, -0.2864592671394348, 0.017265573143959045, 0.08406311273574829, 0.0027674848679453135, 0.06290827691555023, 0.07236549258232117, -0.07389893382787704, 0.11328595131635666, -0.08021481335163116, 0.13019037246704102, 0.08625296503305435, -0.062064990401268005, -0.23071379959583282, -0.07525765895843506, 0.0963398814201355, 0.12251301854848862, 0.06215599179267883, -0.022921854630112648, 0.15455181896686554, -0.06248689442873001, 0.012971068732440472, 0.1294165402650833, -0.11526761949062347, -0.05572471022605896, 0.061741601675748825, 0.11775490641593933, 0.10740239918231964, -0.14110268652439117, -0.0017287094378843904, 0.04900608956813812, 0.029121357947587967, 0.08589313924312592, 0.022661056369543076, 0.12003941088914871, 0.04652795568108559, -0.13695219159126282, -0.04037507623434067, 0.12011898308992386, 0.038862764835357666, -0.06446044892072678, -0.2168138176202774, -0.006778308190405369, -0.0601806715130806, -0.014732478186488152, -0.07019448280334473, 0.039128515869379044, -0.02470310963690281, 0.07317749410867691, -0.04465159401297569, -0.1063927412033081, -0.0421026237308979, 0.0892222449183464, 0.07748593389987946, 0.011527054943144321, -0.02519804798066616, 0.04627908393740654, 0.13455867767333984, 0.05402068421244621, -0.10399353504180908, -0.07017925381660461, -0.06942764669656754, -0.09420394152402878, -0.04035796597599983, 0.056760527193546295, 0.031942449510097504, 0.02665667235851288, 0.22703726589679718, 0.016653569415211678, 0.04155244305729866, 0.0224777739495039, 0.01032855175435543, 0.043662428855895996, 0.0955500528216362, -0.05303520709276199, -0.15660029649734497, -0.04072032496333122, 0.09077946096658707, -0.0027527001220732927, -0.036689214408397675, -0.03966725245118141, 0.03849169611930847, 0.06843466311693192, 0.13122352957725525, 0.07552056759595871, -0.017929591238498688, -0.04813180863857269, -0.030096933245658875, 0.23523783683776855, -0.1493375599384308, 0.04426715523004532, -0.02271856553852558, -0.01804111897945404, -0.03908449783921242, 0.03597262129187584, 0.022118929773569107, -0.000004518366949923802, 0.09706240892410278, -0.058981191366910934, -0.05378659814596176, -0.10168042778968811, -0.03272576630115509, 0.04088849574327469, -0.013975566253066063, -0.010589460842311382, -0.09025166928768158, -0.09490354359149933, -0.04766594246029854, 0.05537205561995506, -0.05123869329690933, -0.03770573064684868, 0.009465423412621021, -0.08151785284280777, -0.005444355774670839, -0.005417742300778627, 0.10699385404586792, -0.03222226724028587, 0.04445803165435791, -0.027600755915045738, 0.05225523188710213, 0.09919606149196625, 0.031576547771692276, -0.0773419588804245, 0.0561848059296608, -0.22559374570846558, 0.07503069192171097, -0.11481974273920059, 0.04335082694888115, -0.1704932004213333, -0.042439818382263184, 0.005444696638733149, 0.0139949731528759, 0.013206101022660732, 0.12720820307731628, -0.19255615770816803, -0.01654396951198578, 0.13260798156261444, -0.09212633967399597, -0.118110790848732, 0.07884611934423447, -0.029701577499508858, 0.1624738723039627, 0.04682036489248276, -0.027025915682315826, 0.09224298596382141, -0.16434773802757263, -0.07092688232660294, -0.00949116237461567, -0.01727987825870514, 0.12109188735485077, 0.07512219995260239, -0.05991523340344429, 0.046571120619773865, 0.02832140028476715, -0.038078423589468, -0.04424772411584854, -0.050857074558734894, -0.10884185880422592, -0.01070026308298111, -0.08987759798765182, 0.04065500199794769, -0.01250192429870367, -0.07916021347045898, -0.029885273426771164, -0.18612512946128845, -0.0030564051121473312, 0.10038342326879501, 0.0035033065360039473, -0.005652366206049919, -0.08666291832923889, 0.026358824223279953, -0.03112892620265484, -0.008404186926782131, -0.16764774918556213, -0.04399421438574791, 0.046902090311050415, -0.16094985604286194, 0.020117372274398804, -0.06413903087377548, 0.06334125250577927, 0.03641495108604431, -0.05590536445379257, -0.0248766727745533, -0.01730942726135254, 0.011945613659918308, -0.05083848536014557, -0.18994836509227753, -0.056277405470609665, -0.037882111966609955, 0.149809330701828, -0.25956398248672485, 0.032966937869787216, 0.051140617579221725, 0.14649195969104767, 0.00406361510977149, -0.05115427449345589, 0.01429014839231968, -0.05360214412212372, -0.054652128368616104, -0.06746816635131836, -0.006135428790003061, -0.027576493099331856, -0.05147203803062439, 0.019243421033024788, -0.1755700707435608, -0.021410830318927765, 0.09424154460430145, 0.12876708805561066, -0.1486445665359497, -0.018640631809830666, -0.048725154250860214, -0.06339836865663528, -0.0715010017156601, -0.07038594037294388, 0.10712739825248718, 0.0513901449739933, 0.04796046018600464, -0.07435787469148636, -0.07092321664094925, 0.02726263552904129, 0.006906150374561548, -0.03382374346256256, 0.08727246522903442, 0.05199531093239784, -0.09209315478801727, 0.0756213590502739, 0.1092359870672226, 0.07177663594484329, 0.09363535046577454, 0.01574566215276718, -0.11756632477045059, -0.028492970392107964, 0.036266472190618515, 0.02740776725113392, 0.1465986967086792, -0.05952361226081848, 0.04016614332795143, 0.04494241625070572, -0.04170418903231621, 0.022319864481687546, -0.08787637203931808, 0.024075502529740334, 0.025203049182891846, -0.0034381982404738665, 0.06284574419260025, -0.02525499276816845, -0.0050758360885083675, 0.07016654312610626, 0.047779910266399384, 0.04621000960469246, 0.009655474685132504, -0.01720241829752922, -0.1047825813293457, 0.16950392723083496, -0.0951867327094078, -0.269941508769989, -0.17632324993610382, 0.026197833940386772, 0.04035249724984169, -0.022378476336598396, 0.031619444489479065, -0.07056326419115067, -0.10630585998296738, -0.1060405746102333, -0.002429972169920802, 0.01714223250746727, -0.06364088505506516, -0.0741225928068161, 0.07348573952913284, 0.04382912442088127, -0.14902326464653015, 0.038552410900592804, 0.055694397538900375, -0.057955220341682434, -0.0233661737293005, 0.09118817001581192, 0.12397737801074982, 0.14583967626094818, -0.021366750821471214, -0.028626007959246635, 0.029004426673054695, 0.19620531797409058, -0.13469526171684265, 0.10371150821447372, 0.13814030587673187, -0.04545360431075096, 0.08360563963651657, 0.1560150384902954, 0.029186224564909935, -0.08317049592733383, 0.05044832453131676, 0.04082648828625679, -0.043159641325473785, -0.2666129767894745, -0.0534592866897583, 0.012832709588110447, -0.06255637854337692, 0.09786593168973923, 0.10183793306350708, 0.11542957276105881, 0.034910861402750015, -0.07166364789009094, -0.043925940990448, -0.0058974819257855415, 0.11737963557243347, -0.05490213260054588, -0.012639665976166725, 0.07686592638492584, -0.05086168646812439, 0.005355054512619972, 0.10266812145709991, 0.02973790094256401, 0.17442677915096283, 0.020399179309606552, 0.11231429129838943, 0.06195578724145889, 0.08633565157651901, 0.0007386076031252742, 0.02951662428677082, 0.05147615820169449, 0.017203815281391144, -0.002300140680745244, -0.10421168059110641, -0.006156572140753269, 0.1449710875749588, 0.028103826567530632, 0.029669636860489845, -0.0018948549404740334, -0.005003341939300299, 0.05121048167347908, 0.1746254414319992, -0.011592294089496136, -0.22072425484657288, -0.0845772922039032, 0.06936841458082199, -0.06218599155545235, -0.12968985736370087, -0.026130788028240204, 0.045467354357242584, -0.17519839107990265, 0.026703642681241035, -0.027433741837739944, 0.0919293761253357, -0.09345759451389313, -0.02221956104040146, 0.03687324374914169, 0.084866963326931, -0.014529162086546421, 0.08703910559415817, -0.14498743414878845, 0.11886418610811234, 0.02978132851421833, 0.09024628251791, -0.11081171780824661, 0.07909037172794342, -0.007550720125436783, 0.009180475026369095, 0.19379350543022156, -0.011335089802742004, -0.03514958545565605, -0.08774717897176743, -0.11210042238235474, -0.013537433929741383, 0.12687496840953827, -0.1243172138929367, 0.08773399889469147, -0.015198243781924248, -0.044079482555389404, 0.00937260314822197, -0.12100647389888763, -0.17273177206516266, -0.19628387689590454, 0.05585884302854538, -0.09575839340686798, 0.025643249973654747, -0.11914430558681488, -0.07089093327522278, -0.02952558360993862, 0.241120383143425, -0.1745356321334839, -0.06510113179683685, -0.1468164622783661, -0.046294767409563065, 0.1662203073501587, -0.04437198117375374, 0.0718095526099205, -0.0208172257989645, 0.20345525443553925, 0.005988610442727804, -0.004939318168908358, 0.06724198162555695, -0.08892562240362167, -0.16873881220817566, -0.06771010160446167, 0.1510489284992218, 0.11680185794830322, 0.04907919466495514, -0.002248800592496991, 0.0011772146681323647, -0.016943959519267082, -0.1137804463505745, -0.0033210667315870523, 0.16037839651107788, 0.03878779336810112, 0.025986969470977783, -0.05243593826889992, -0.08797456324100494, -0.06899320334196091, -0.06853509694337845, 0.06221301481127739, 0.19590823352336884, -0.10376439243555069, 0.1700313836336136, 0.147536963224411, -0.07305635511875153, -0.23175598680973053, 0.035342130810022354, 0.04983805492520332, 0.0014306638622656465, 0.04886869341135025, -0.18252557516098022, 0.10521943867206573, 0.019543392583727837, -0.05505957826972008, 0.13485197722911835, -0.1557481735944748, -0.1552847921848297, 0.0722852572798729, 0.03904085233807564, -0.22423844039440155, -0.1354004591703415, -0.09622503817081451, -0.05825018882751465, -0.14065024256706238, 0.06054598465561867, -0.002136280992999673, 0.015948504209518433, 0.03500790148973465, -0.0015643214574083686, 0.027123261243104935, -0.058935679495334625, 0.18609118461608887, -0.004065449349582195, 0.020676052197813988, -0.060264769941568375, -0.0478842556476593, 0.09839435666799545, -0.06130504235625267, 0.12208222597837448, 0.004057085141539574, 0.01594383642077446, -0.10362856835126877, -0.048314861953258514, -0.04328322783112526, 0.05154227837920189, -0.07548051327466965, -0.10070807486772537, -0.043625857681035995, 0.08841723203659058, 0.07005169242620468, -0.03383097052574158, 0.00549331633374095, -0.07189501076936722, 0.10019614547491074, 0.17795267701148987, 0.17573626339435577, 0.009926567785441875, -0.07241068035364151, 0.01677953451871872, -0.04142116755247116, 0.044231921434402466, -0.2513144314289093, 0.03756171092391014, 0.06098250672221184, 0.029438555240631104, 0.09217222779989243, -0.020435843616724014, -0.1820858269929886, -0.04050002992153168, 0.08094815909862518, -0.05452597141265869, -0.22617179155349731, -0.019085140898823738, 0.0954197570681572, -0.2020406424999237, -0.007372708059847355, 0.03995226323604584, -0.048725228756666183, -0.023169852793216705, 0.00010950004070764408, 0.06317184865474701, 0.002471912419423461, 0.09773622453212738, 0.0735151618719101, 0.09715340286493301, -0.08337292820215225, 0.10562895983457565, 0.10150538384914398, -0.09572599828243256, 0.03605884686112404, 0.06754924356937408, -0.05300498008728027, -0.043293699622154236, 0.03665391728281975, 0.033023297786712646, 0.005234600510448217, -0.060321882367134094, 0.013913018628954887, -0.036497246474027634, 0.044923391193151474, 0.08326134830713272, 0.03754979372024536, -0.013354414142668247, 0.06462216377258301, 0.03401726484298706, -0.10898099094629288, 0.10366570204496384, 0.01731540448963642, 0.04105307161808014, -0.08384523540735245, -0.019968897104263306, 0.035425446927547455, 0.030576206743717194, -0.01765924133360386, -0.02306121215224266, -0.02860277332365513, -0.01614218018949032, -0.14299540221691132, -0.023106401786208153, -0.07243485748767853, 0.006181265693157911, 0.014656842686235905, -0.031884219497442245, -0.011233693920075893, 0.02475680410861969, -0.06979699432849884, -0.07426341623067856, -0.006949664559215307, 0.09833318740129471, -0.15115703642368317, 0.008848577737808228, 0.06907843053340912, -0.11088496446609497, 0.08190931379795074, -0.008411259390413761, 0.016245156526565552, 0.022527478635311127, -0.15448406338691711, 0.05601610988378525, 0.0008648968650959432, 0.01916889287531376, 0.025886621326208115, -0.16471809148788452, 0.004104440100491047, -0.04661374166607857, -0.02149827405810356, -0.00004464812809601426, -0.02647159807384014, -0.12325995415449142, 0.06858719140291214, -0.015622655861079693, -0.035931166261434555, -0.02701525390148163, 0.0539589487016201, 0.07888586074113846, -0.027474910020828247, 0.10445091128349304, -0.008690856397151947, 0.04941811040043831, -0.16801609098911285, -0.02470702864229679, -0.04982255399227142, 0.019377702847123146, 0.009884213097393513, -0.007693959400057793, 0.04183054715394974, -0.00976533442735672, 0.21883612871170044, -0.05075952783226967, 0.1607085019350052, 0.05847611650824547, -0.017352959141135216, -0.0007513365126214921, 0.06180921941995621, 0.05997028574347496, 0.04658793285489082, 0.009480604901909828, 0.023740366101264954, -0.022450892254710197, -0.006695089396089315, -0.15932634472846985, 0.01890849508345127, 0.14999441802501678, 0.06301083415746689, 0.024745315313339233, 0.05866100639104843, -0.12775006890296936, -0.12135478109121323, 0.09311001747846603, -0.026755332946777344, 0.00928465835750103, -0.08245618641376495, 0.1358020007610321, 0.14980104565620422, -0.14000412821769714, 0.05256148427724838, -0.06134212389588356, -0.05217423290014267, -0.10388828068971634, -0.12032219022512436, -0.05887215584516525, -0.053666237741708755, 0.002330566756427288, -0.03760887682437897, 0.054546963423490524, 0.03344334661960602, -0.009351172484457493, -0.00022941511997487396, 0.13597318530082703, -0.019751882180571556, -0.0028988157864660025, 0.048313532024621964, 0.03693558648228645, 0.02373051457107067, -0.05275435373187065, 0.02940409444272518, 0.02539868652820587, 0.032232340425252914, 0.06546790152788162, 0.033412106335163116, -0.047448933124542236, 0.03804153576493263, -0.0025254099164158106, -0.11207924783229828, 0.019641218706965446, -0.00460948096588254, -0.0742158442735672, 0.1268945336341858, 0.0407399944961071, 0.010224059224128723, -0.03741471841931343, 0.24361543357372284, -0.06653323769569397, -0.06378097087144852, -0.13251738250255585, 0.10491154342889786, -0.0027236645109951496, 0.06476365029811859, 0.023412218317389488, -0.1284150779247284, 0.005243356805294752, 0.13858191668987274, 0.12181595712900162, 0.0045748427510261536, 0.009228081442415714, 0.0518609918653965, 0.0025186820421367884, -0.06998204439878464, 0.054019294679164886, 0.06992026418447495, 0.12919506430625916, -0.07847554981708527, 0.07680778950452805, 0.0006860480643808842, -0.08370215445756912, -0.02947772853076458, 0.11312682181596756, -0.0409729965031147, 0.03491825982928276, -0.047444481402635574, 0.10916327685117722, -0.05787910893559456, -0.29412412643432617, 0.02350960113108158, -0.09588567912578583, -0.15202060341835022, -0.018367812037467957, 0.05944539234042168, -0.02624768204987049, 0.018029648810625076, 0.06971040368080139, -0.06011629104614258, 0.20098382234573364, 0.0335683599114418, -0.07864278554916382, -0.0664360448718071, 0.04837050288915634, -0.06564252078533173, 0.2949807047843933, 0.008418165147304535, 0.02863333560526371, 0.10770907253026962, -0.03253700211644173, -0.18271861970424652, 0.010723991319537163, 0.1133992001414299, -0.08056149631738663, 0.08200647681951523, 0.19000613689422607, -0.012578671798110008, 0.1209007054567337, 0.05294662341475487, -0.047376248985528946, 0.04217283055186272, -0.03389401361346245, -0.051268599927425385, -0.10752558708190918, 0.058453381061553955, -0.05909625440835953, 0.15447644889354706, 0.10152646154165268, -0.05671518296003342, -0.004550917539745569, -0.05555408447980881, 0.04875178262591362, 0.01804669201374054, 0.12263146042823792, 0.02951994352042675, -0.1865430772304535, 0.032826557755470276, -0.01144319772720337, 0.10186848044395447, -0.25588861107826233, -0.08421015739440918, 0.08833149075508118, -0.011924264021217823, -0.05105875805020332, 0.10560628771781921, 0.057650718837976456, 0.04243382066488266, -0.043439045548439026, -0.10480839014053345, -0.02186836116015911, 0.14663739502429962, -0.1469624787569046, -0.025013303384184837 ]
null
null
peft
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # CaseSummarization_4.0 This model is a fine-tuned version of [TheBloke/zephyr-7B-beta-GPTQ](https://huggingface.co/TheBloke/zephyr-7B-beta-GPTQ) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - training_steps: 250 - mixed_precision_training: Native AMP ### Training results ### Framework versions - PEFT 0.8.2 - Transformers 4.37.2 - Pytorch 2.1.2 - Datasets 2.17.0 - Tokenizers 0.15.1
{"license": "mit", "library_name": "peft", "tags": ["trl", "sft", "generated_from_trainer"], "base_model": "TheBloke/zephyr-7B-beta-GPTQ", "model-index": [{"name": "CaseSummarization_4.0", "results": []}]}
null
Bhagya17/CaseSummarization_4.0
[ "peft", "safetensors", "mistral", "trl", "sft", "generated_from_trainer", "base_model:TheBloke/zephyr-7B-beta-GPTQ", "license:mit", "4-bit", "region:us" ]
2024-02-10T13:32:08+00:00
[]
[]
TAGS #peft #safetensors #mistral #trl #sft #generated_from_trainer #base_model-TheBloke/zephyr-7B-beta-GPTQ #license-mit #4-bit #region-us
# CaseSummarization_4.0 This model is a fine-tuned version of TheBloke/zephyr-7B-beta-GPTQ on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0002 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - training_steps: 250 - mixed_precision_training: Native AMP ### Training results ### Framework versions - PEFT 0.8.2 - Transformers 4.37.2 - Pytorch 2.1.2 - Datasets 2.17.0 - Tokenizers 0.15.1
[ "# CaseSummarization_4.0\n\nThis model is a fine-tuned version of TheBloke/zephyr-7B-beta-GPTQ on the None dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- training_steps: 250\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.2\n- Datasets 2.17.0\n- Tokenizers 0.15.1" ]
[ "TAGS\n#peft #safetensors #mistral #trl #sft #generated_from_trainer #base_model-TheBloke/zephyr-7B-beta-GPTQ #license-mit #4-bit #region-us \n", "# CaseSummarization_4.0\n\nThis model is a fine-tuned version of TheBloke/zephyr-7B-beta-GPTQ on the None dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- training_steps: 250\n- mixed_precision_training: Native AMP", "### Training results", "### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.2\n- Datasets 2.17.0\n- Tokenizers 0.15.1" ]
[ 57, 39, 6, 12, 8, 3, 102, 4, 36 ]
[ "passage: TAGS\n#peft #safetensors #mistral #trl #sft #generated_from_trainer #base_model-TheBloke/zephyr-7B-beta-GPTQ #license-mit #4-bit #region-us \n# CaseSummarization_4.0\n\nThis model is a fine-tuned version of TheBloke/zephyr-7B-beta-GPTQ on the None dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0002\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- training_steps: 250\n- mixed_precision_training: Native AMP### Training results### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.37.2\n- Pytorch 2.1.2\n- Datasets 2.17.0\n- Tokenizers 0.15.1" ]
[ -0.10381967574357986, 0.04290097951889038, -0.0022309201303869486, 0.07671438157558441, 0.10759361833333969, 0.033689260482788086, 0.1297806352376938, 0.11944634467363358, -0.0675017386674881, 0.061892397701740265, 0.05110137537121773, 0.0196120273321867, 0.06115921959280968, 0.18580974638462067, -0.016543086618185043, -0.2385607659816742, 0.013365834951400757, -0.011181922629475594, -0.0438990518450737, 0.11183357983827591, 0.10853004455566406, -0.11376991122961044, 0.042867261916399, -0.0004157815710641444, -0.1361629217863083, 0.0033449153415858746, -0.014672048389911652, -0.02796294540166855, 0.10787879675626755, 0.012097708880901337, 0.11373073607683182, 0.024675985798239708, 0.1585591733455658, -0.2320335954427719, 0.009301350452005863, 0.0858735665678978, 0.04746803641319275, 0.06967775523662567, 0.04946882277727127, 0.002889919327571988, 0.09047740697860718, -0.14978986978530884, 0.09584472328424454, 0.01456616073846817, -0.1035846546292305, -0.1957341730594635, -0.11560243368148804, 0.0789945051074028, 0.10445421934127808, 0.077321857213974, 0.0006811984349042177, 0.14547574520111084, -0.09720629453659058, 0.04523046687245369, 0.2307145744562149, -0.2484653741121292, -0.0923953652381897, 0.07797826081514359, 0.0852922722697258, 0.06324969977140427, -0.12824037671089172, -0.02485872432589531, 0.042610589414834976, 0.02845439873635769, 0.08236557245254517, -0.022187914699316025, -0.06463910639286041, -0.014551205560564995, -0.13806100189685822, -0.005779471714049578, 0.11057333648204803, 0.024336248636245728, -0.049216169863939285, -0.09146371483802795, -0.02312612347304821, -0.08044856786727905, -0.0326412208378315, -0.026237420737743378, 0.03401758149266243, -0.03212788328528404, -0.02356082946062088, -0.04942089319229126, -0.08728855848312378, -0.09653674811124802, 0.001404282753355801, 0.1218915656208992, 0.02607867866754532, 0.002763410797342658, -0.017692772671580315, 0.13105815649032593, -0.027699820697307587, -0.09248476475477219, -0.017245998606085777, -0.012361959554255009, -0.09454885870218277, -0.07687442004680634, -0.029085833579301834, 0.011209111660718918, 0.023540815338492393, 0.15288735926151276, -0.09313613921403885, 0.08649809658527374, 0.008207105100154877, 0.024786073714494705, -0.04559706524014473, 0.0763765498995781, -0.007436023093760014, -0.030552180483937263, -0.0015382617712020874, 0.0997871682047844, -0.013472053222358227, 0.001961776753887534, -0.06340967118740082, -0.03596622496843338, 0.05053267627954483, 0.06454335153102875, -0.06397014111280441, 0.031269509345293045, -0.05823758617043495, -0.014172063209116459, 0.037611767649650574, -0.10360292345285416, 0.04238271340727806, 0.003628499573096633, -0.041773825883865356, -0.06261175870895386, 0.003161267377436161, 0.021100085228681564, 0.002428076695650816, 0.10021563619375229, -0.06037764623761177, 0.00573117146268487, -0.09031178802251816, -0.06987930834293365, 0.01938316412270069, -0.037325698882341385, -0.024364663287997246, -0.07379025965929031, -0.20680242776870728, -0.0683097243309021, 0.022350475192070007, -0.05112898349761963, -0.010300171561539173, -0.046754200011491776, -0.0348934568464756, 0.023283012211322784, -0.03383290395140648, 0.16521291434764862, -0.06105787679553032, 0.08473251014947891, -0.05664810165762901, 0.016925690695643425, -0.014150322414934635, 0.006315909791737795, -0.08248061686754227, 0.03791109472513199, -0.13114005327224731, 0.05156942456960678, -0.10398129373788834, -0.014052518643438816, -0.1295710951089859, -0.09607137739658356, -0.033143624663352966, -0.02353307604789734, 0.0766558051109314, 0.11102025210857391, -0.20604431629180908, -0.02055092342197895, 0.1789921373128891, -0.09849169105291367, -0.04275066778063774, 0.08015439659357071, -0.062605120241642, 0.056083161383867264, 0.06222854554653168, 0.17356853187084198, 0.11839562654495239, -0.14435526728630066, 0.029143521562218666, -0.025555016472935677, 0.08271155506372452, 0.06081452965736389, 0.054094985127449036, -0.034464314579963684, -0.020997248589992523, -0.0021705625113099813, -0.09184572845697403, 0.011829005554318428, -0.08498186618089676, -0.06407568603754044, -0.032108861953020096, -0.07262390851974487, 0.0806511789560318, 0.02071031741797924, 0.01217008475214243, -0.0947912260890007, -0.08341610431671143, 0.105995774269104, 0.13413137197494507, -0.051235489547252655, 0.003244927152991295, -0.056577179580926895, 0.009182567708194256, 0.012662983499467373, -0.03185100853443146, -0.1687474548816681, -0.11843141168355942, 0.0339500866830349, -0.06702392548322678, 0.03004639968276024, 0.014979173429310322, 0.07357090711593628, 0.06457823514938354, -0.08790779113769531, -0.00801905058324337, -0.09205366671085358, 0.01584937423467636, -0.11713273078203201, -0.20799316465854645, -0.04343177378177643, -0.02613561600446701, 0.17309105396270752, -0.24454368650913239, 0.015960512682795525, 0.004776042886078358, 0.14273761212825775, 0.028460878878831863, -0.054566435515880585, -0.01350573729723692, 0.07772015035152435, 0.02821458876132965, -0.09053774178028107, 0.04953619837760925, 0.0016761342994868755, -0.0750577300786972, -0.0246239323168993, -0.14745861291885376, 0.005405232775956392, 0.06967424601316452, 0.07966487109661102, -0.12070655822753906, -0.1232132762670517, -0.0627705454826355, -0.035860735923051834, -0.10186916589736938, 0.007516445592045784, 0.1746603399515152, 0.0038266207557171583, 0.10151789337396622, -0.07107463479042053, -0.06888440251350403, -0.003586570965126157, -0.004050415009260178, 0.054105181246995926, 0.08215799927711487, 0.05859391763806343, -0.13674449920654297, 0.090911865234375, 0.11156672984361649, -0.05006953701376915, 0.14806964993476868, -0.05524936690926552, -0.0757279098033905, -0.016918541863560677, 0.024065416306257248, -0.005278768017888069, 0.1495221108198166, -0.04014907404780388, 0.02924300730228424, 0.015731964260339737, 0.039680369198322296, 0.01622656360268593, -0.1938926726579666, -0.01366820465773344, 0.003432946279644966, -0.04355735331773758, -0.014285736717283726, -0.022343328222632408, 0.03770133852958679, 0.10148035734891891, 0.017203310504555702, 0.0037670950405299664, 0.00942822266370058, -0.010823623277246952, -0.09718110412359238, 0.18979567289352417, -0.13846364617347717, -0.13184663653373718, -0.07768207788467407, 0.08257679641246796, 0.0071897609159350395, -0.03052489273250103, 0.011864891275763512, -0.07612472027540207, -0.021343084052205086, -0.08251584321260452, 0.0009789754403755069, -0.012238144874572754, -0.015484808012843132, -0.004433430265635252, 0.021818095818161964, 0.06889130175113678, -0.11567381024360657, 0.0008899283711798489, -0.018486110493540764, -0.08792124688625336, 0.02524024248123169, 0.03306153044104576, 0.06621825695037842, 0.1334896683692932, -0.019994867965579033, -0.006988827139139175, -0.07010766863822937, 0.19248715043067932, -0.09251275658607483, 0.007674480322748423, 0.10063866525888443, 0.0011076298542320728, 0.03459380194544792, 0.09816127270460129, 0.03876886144280434, -0.10565973818302155, 0.037065908312797546, 0.06720609217882156, -0.021725833415985107, -0.22947874665260315, -0.05066579207777977, -0.04941021651029587, -0.07805673778057098, 0.10607694089412689, 0.051656294614076614, -0.018009765073657036, 0.04241431504487991, -0.025612149387598038, 0.01636754348874092, 0.03483268991112709, 0.08684104681015015, 0.07918677479028702, 0.037958864122629166, 0.10046670585870743, -0.026550939306616783, -0.0105147585272789, 0.07000948488712311, 0.030955541878938675, 0.2576671242713928, -0.013329174369573593, 0.035817377269268036, 0.07855889946222305, 0.14375703036785126, -0.0006778359529562294, 0.021347351372241974, 0.044963289052248, -0.013604976236820221, -0.012015913613140583, -0.07131335884332657, -0.026351207867264748, 0.036868780851364136, -0.024931153282523155, 0.025543317198753357, -0.08293833583593369, -0.047441158443689346, 0.03155610337853432, 0.2577452063560486, 0.045852214097976685, -0.2465640753507614, -0.07542276382446289, 0.016072673723101616, -0.03356518968939781, -0.06722626090049744, -0.0002118818083545193, 0.13755834102630615, -0.1508507877588272, 0.0709649920463562, -0.05584600940346718, 0.08913900703191757, -0.029986493289470673, -0.0064989617094397545, 0.0438883900642395, 0.10163362324237823, -0.024051038548350334, 0.06413181871175766, -0.2364177107810974, 0.23757676780223846, 0.01589122600853443, 0.10916716605424881, -0.05689968913793564, 0.009538229554891586, 0.03634185716509819, 0.05963287502527237, 0.09721041470766068, 0.004904695320874453, -0.0938459262251854, -0.19009529054164886, -0.046954017132520676, 0.04405510425567627, 0.12499548494815826, -0.046153467148542404, 0.06975502520799637, -0.037134796380996704, 0.023026427254080772, 0.03941497206687927, -0.07485038787126541, -0.20200751721858978, -0.10020957142114639, 0.0037085891235619783, 0.014970351941883564, -0.025698568671941757, -0.12005853652954102, -0.09017282724380493, -0.01089791115373373, 0.13858893513679504, -0.021949484944343567, -0.03140230104327202, -0.14123299717903137, 0.08036479353904724, 0.1227063313126564, -0.03302548825740814, 0.028280984610319138, 0.0460757315158844, 0.11914567649364471, 0.002032431773841381, -0.04846412315964699, 0.06368670612573624, -0.0793752521276474, -0.21859100461006165, -0.08855137228965759, 0.13623405992984772, 0.09755665063858032, 0.05894261226058006, 0.013783968985080719, 0.04141610115766525, 0.04116014763712883, -0.10243942588567734, 0.009000170975923538, 0.1260591298341751, 0.07440425455570221, 0.05538822337985039, -0.07649069279432297, 0.04071252420544624, -0.02642062120139599, -0.04903663322329521, 0.09240647405385971, 0.2533208131790161, -0.0891704112291336, 0.08921892940998077, 0.04958599433302879, -0.09706743061542511, -0.18497438728809357, 0.09549368172883987, 0.12998472154140472, 0.028726305812597275, 0.051631588488817215, -0.1684451699256897, 0.06761515885591507, 0.14644837379455566, -0.028252698481082916, 0.05579685419797897, -0.3146193027496338, -0.12952697277069092, 0.05942099913954735, 0.08994164317846298, -0.012445369735360146, -0.11991620063781738, -0.03455200418829918, -0.031042814254760742, -0.10778415948152542, 0.10139638930559158, -0.13852521777153015, 0.10157359391450882, -0.0016387333162128925, 0.041434429585933685, 0.02537553571164608, -0.028109654784202576, 0.16380226612091064, -0.003444963600486517, 0.09787175059318542, -0.04205719754099846, 0.04723755642771721, 0.05399956554174423, -0.05749678984284401, 0.005281778052449226, -0.021959587931632996, 0.05514930933713913, -0.09822237491607666, -0.030155861750245094, -0.07860863953828812, 0.06546499580144882, -0.04355909675359726, -0.06964549422264099, -0.06698343902826309, 0.08462049067020416, 0.02770981751382351, -0.035939306020736694, 0.013937942683696747, -0.034635335206985474, 0.150129497051239, 0.0503322072327137, 0.12396200746297836, 0.02590331621468067, -0.07361263036727905, 0.013580613769590855, -0.039444804191589355, 0.06372784078121185, -0.10925596207380295, 0.01892617717385292, 0.13063639402389526, 0.04553768038749695, 0.13889849185943604, 0.03990480303764343, -0.060571447014808655, 0.009845038875937462, 0.04595816135406494, -0.08381468802690506, -0.12138400226831436, 0.02417692355811596, 0.05923959240317345, -0.1196175068616867, -0.0028132488951087, 0.11792764812707901, -0.0677545964717865, -0.02929193712770939, -0.00935981422662735, 0.016963543370366096, -0.0523982048034668, 0.21058380603790283, 0.027205975726246834, 0.06683146953582764, -0.07660539448261261, 0.0979333370923996, 0.05968502163887024, -0.051351480185985565, 0.03292402625083923, 0.06803223490715027, -0.09539869427680969, -0.01952526345849037, 0.09875594824552536, 0.13422667980194092, -0.004608842544257641, -0.0463981032371521, -0.05846932530403137, -0.1137409582734108, 0.03142556548118591, 0.13269877433776855, 0.03279099240899086, -0.0059042624197900295, 0.0016641392139717937, 0.03674069792032242, -0.10215841233730316, 0.05516685172915459, 0.03196786344051361, 0.07948591560125351, -0.08719320595264435, 0.15447764098644257, 0.012254219502210617, 0.0022208045702427626, -0.013464296236634254, 0.036089763045310974, -0.11537424474954605, -0.00011587636254262179, -0.11400633305311203, 0.004685171414166689, -0.03502124547958374, 0.00316701945848763, 0.0021229255944490433, -0.04194661229848862, -0.013074718415737152, 0.037103321403265, -0.07809006422758102, -0.04153547063469887, -0.000638102472294122, 0.07295708358287811, -0.10752425342798233, -0.02090948075056076, 0.045384205877780914, -0.07807055115699768, 0.059091877192258835, 0.020237641409039497, 0.0403415746986866, 0.05537380650639534, -0.17464004456996918, 0.00675445981323719, 0.03136667609214783, 0.020095836371183395, 0.02395755797624588, -0.0935734361410141, -0.01353182177990675, -0.03239692002534866, 0.050633255392313004, 0.04269903153181076, 0.023166997358202934, -0.10415590554475784, -0.042257051914930344, -0.05085292458534241, -0.06738467514514923, -0.03784580156207085, 0.02916664443910122, 0.08264622837305069, 0.0508914589881897, 0.12076582759618759, -0.10325915366411209, 0.045115143060684204, -0.19625508785247803, -0.04592732712626457, -0.019104423001408577, -0.02139011025428772, -0.07502931356430054, -0.043427396565675735, 0.09955892711877823, -0.02973128855228424, 0.11634952574968338, -0.013705495744943619, 0.07073089480400085, 0.031168054789304733, -0.08457986265420914, -0.0017335452139377594, 0.02570982277393341, 0.19562707841396332, 0.053857043385505676, -0.022133463993668556, 0.05754135176539421, 0.002216321649029851, 0.02224336564540863, 0.053739625960588455, 0.205588236451149, 0.17791585624217987, 0.01596527360379696, 0.03768065571784973, 0.04898060858249664, -0.11811507493257523, -0.10321890562772751, 0.10233298689126968, -0.0028172757010906935, 0.07073283195495605, -0.08779484033584595, 0.19366271793842316, 0.08077280968427658, -0.2075696587562561, 0.044222839176654816, -0.07782626152038574, -0.09867240488529205, -0.1123846173286438, -0.04239924997091293, -0.07213094085454941, -0.1296623945236206, -0.0011733687715604901, -0.11036772280931473, 0.03380421921610832, 0.0955338105559349, 0.01933160237967968, 0.03966203331947327, 0.1355501115322113, -0.027165591716766357, 0.0029380626510828733, 0.06652146577835083, 0.01921851933002472, 0.02747814543545246, -0.10435086488723755, -0.08749423921108246, 0.06638030707836151, -0.037548378109931946, 0.06187134608626366, -0.060128308832645416, -0.0022608533035963774, 0.01021733321249485, 0.011763363145291805, -0.06035921722650528, 0.03984272480010986, 0.013116213493049145, 0.04638131335377693, 0.07519707828760147, 0.06622463464736938, 0.0005520321428775787, -0.03177696466445923, 0.2728552520275116, -0.057065773755311966, -0.1162908673286438, -0.13035818934440613, 0.2456706017255783, 0.005013978574424982, -0.007524306420236826, 0.04689585044980049, -0.09757345914840698, 0.038723330944776535, 0.13331107795238495, 0.1331646591424942, -0.057349417358636856, 0.012922437861561775, 0.0024375859647989273, -0.023904483765363693, -0.0610700361430645, 0.14421014487743378, 0.07527932524681091, 0.006527397781610489, -0.07161619514226913, 0.004684389568865299, 0.0024978844448924065, -0.01675672084093094, -0.0489237904548645, 0.037964727729558945, 0.007900137454271317, 0.004288812633603811, -0.04847748950123787, 0.10265952348709106, 0.03204081207513809, -0.16650930047035217, 0.07317537069320679, -0.13372434675693512, -0.1700548529624939, -0.010162731632590294, 0.022471288219094276, -0.006846414878964424, 0.02558157593011856, -0.03446120023727417, 0.00970774982124567, 0.11054452508687973, -0.026936357840895653, -0.003916177898645401, -0.15198276937007904, 0.09389737248420715, -0.03787544369697571, 0.224668487906456, -0.02278490737080574, 0.08201467990875244, 0.09576144069433212, 0.02593120001256466, -0.11315274983644485, 0.05401919409632683, 0.058731261640787125, -0.07461659610271454, 0.012456631287932396, 0.12658186256885529, -0.048736680299043655, 0.10797971487045288, 0.0613623782992363, -0.15048494935035706, 0.024667993187904358, -0.02760692685842514, -0.03273279219865799, -0.095072902739048, 0.034608568996191025, -0.06937477737665176, 0.15350282192230225, 0.1774091124534607, -0.03330552577972412, 0.008322542533278465, -0.05994841456413269, 0.03767099976539612, 0.04280679672956467, 0.11249047517776489, -0.02075621485710144, -0.20042550563812256, 0.023326965048909187, 0.04372991994023323, 0.023748647421598434, -0.20253893733024597, -0.10071278363466263, 0.02766946516931057, -0.054479386657476425, -0.03808305785059929, 0.13007980585098267, 0.029090898111462593, 0.03250729665160179, -0.035002514719963074, -0.2520074248313904, -0.021794447675347328, 0.14331339299678802, -0.11974399536848068, -0.0367724671959877 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text2text-generation
Professor/double-quant-davlan
[ "transformers", "safetensors", "mt5", "text2text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "4-bit", "region:us" ]
2024-02-10T13:37:31+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #mt5 #text2text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #mt5 #text2text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 62, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #mt5 #text2text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #4-bit #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.04424791410565376, 0.17212572693824768, -0.005364083219319582, 0.017688874155282974, 0.10341756790876389, 0.00779188284650445, 0.05365653708577156, 0.11789484322071075, -0.04384218528866768, 0.11921608448028564, 0.043076321482658386, 0.10880826413631439, 0.11680098623037338, 0.15051773190498352, 0.002164856530725956, -0.21935151517391205, 0.04954215884208679, -0.11219757050275803, -0.025547046214342117, 0.12055613845586777, 0.1482943594455719, -0.0999172106385231, 0.07405153661966324, -0.03270019218325615, -0.021699804812669754, -0.03417343273758888, -0.06192147731781006, -0.04243922606110573, 0.03854668140411377, 0.0570434033870697, 0.06276948750019073, 0.004655817523598671, 0.08679546415805817, -0.2714121639728546, 0.018561575561761856, 0.06915108114480972, -0.003790256567299366, 0.0647420734167099, 0.06828988343477249, -0.06163584068417549, 0.10444440692663193, -0.05105012282729149, 0.14006544649600983, 0.08942738175392151, -0.0917397290468216, -0.18559327721595764, -0.09213837236166, 0.1007617712020874, 0.17823244631290436, 0.05264750123023987, -0.024997182190418243, 0.09699685126543045, -0.08634494990110397, 0.017606910318136215, 0.052454520016908646, -0.0865173414349556, -0.053131017833948135, 0.06277287006378174, 0.08982466906309128, 0.05617585778236389, -0.12288065999746323, -0.032797254621982574, 0.0021210145205259323, 0.01897539384663105, 0.07021352648735046, 0.020143844187259674, 0.15232503414154053, 0.03425682336091995, -0.13553716242313385, -0.0494757816195488, 0.10819033533334732, 0.03920786455273628, -0.04256099835038185, -0.24328382313251495, -0.03248966485261917, -0.04018425568938255, -0.03379657864570618, -0.04205594211816788, 0.04236490651965141, -0.001708693103864789, 0.08268597722053528, -0.004922281950712204, -0.07300832122564316, -0.03556794673204422, 0.06502433866262436, 0.06217541545629501, 0.028706608340144157, -0.01722957193851471, 0.017671262845396996, 0.10910706222057343, 0.1033179834485054, -0.11669570952653885, -0.06150662526488304, -0.06709051877260208, -0.07671888917684555, -0.042902201414108276, 0.030323747545480728, 0.01380479708313942, 0.0639425590634346, 0.25972238183021545, 0.021647561341524124, 0.06191372498869896, 0.033041369169950485, 0.007180255837738514, 0.05240538343787193, 0.10798594355583191, -0.06526714563369751, -0.11207113415002823, -0.02147996611893177, 0.08680415898561478, 0.012889539822936058, -0.03865906968712807, -0.051433127373456955, 0.06602397561073303, 0.041762154549360275, 0.11215431243181229, 0.09531506150960922, 0.0205018762499094, -0.0700579360127449, -0.06477990746498108, 0.1923045963048935, -0.15995152294635773, 0.03603306785225868, 0.040534403175115585, -0.03944610804319382, -0.01267994288355112, 0.014780652709305286, 0.021069897338747978, -0.031306613236665726, 0.08206355571746826, -0.05666749179363251, -0.04726353660225868, -0.11216065287590027, -0.03513505309820175, 0.039633043110370636, 0.012883124873042107, -0.03471590578556061, -0.039880767464637756, -0.0757586881518364, -0.08609695732593536, 0.08937917649745941, -0.06941234320402145, -0.05312260612845421, -0.022837279364466667, -0.08434390276670456, 0.023343240842223167, 0.022552719339728355, 0.08065258711576462, -0.022669529542326927, 0.05111207440495491, -0.04905402660369873, 0.058490388095378876, 0.10718686133623123, 0.03864121437072754, -0.057572394609451294, 0.05877876654267311, -0.23786561191082, 0.08956784009933472, -0.06843909621238708, 0.06475556641817093, -0.15738728642463684, -0.022312356159090996, 0.040109679102897644, 0.005456115119159222, -0.005386885721236467, 0.13684538006782532, -0.20837163925170898, -0.025969142094254494, 0.16446895897388458, -0.09622325748205185, -0.07058661431074142, 0.05039175972342491, -0.0434369295835495, 0.10741866379976273, 0.03037932515144348, -0.009797328151762486, 0.06300465017557144, -0.11211669445037842, -0.0045789750292897224, -0.05525997653603554, -0.023780696094036102, 0.13806255161762238, 0.07908279448747635, -0.07613323628902435, 0.06370917707681656, 0.024881809949874878, -0.024640332907438278, -0.056786052882671356, -0.01756562851369381, -0.10223362594842911, 0.0194342490285635, -0.06580051779747009, 0.006488845217972994, -0.018359681591391563, -0.08997378498315811, -0.025971924886107445, -0.17071400582790375, -0.0351765938103199, 0.08182147890329361, -0.0039688036777079105, -0.014921153895556927, -0.11541282385587692, 0.018099943175911903, 0.032839156687259674, 0.006888494826853275, -0.1356445848941803, -0.039948172867298126, 0.03175674006342888, -0.15897881984710693, 0.03711456432938576, -0.06714559346437454, 0.0535677932202816, 0.019491594284772873, -0.02542215958237648, -0.02622687816619873, 0.019500628113746643, 0.008122220635414124, -0.014907474629580975, -0.24132290482521057, -0.03159322217106819, -0.02769308164715767, 0.1711985468864441, -0.1995917558670044, 0.034021735191345215, 0.0848696306347847, 0.152849480509758, 0.006732362788170576, -0.0493493489921093, 0.012774255126714706, -0.07175669819116592, -0.023133842274546623, -0.05594019591808319, 0.0023269858211278915, -0.018618987873196602, -0.04103803262114525, 0.03460358828306198, -0.1767825037240982, -0.04571788012981415, 0.09877919405698776, 0.045814454555511475, -0.13424912095069885, -0.018067775294184685, -0.036899544298648834, -0.053357306867837906, -0.03919341042637825, -0.06328054517507553, 0.10097826272249222, 0.05898883193731308, 0.04317450895905495, -0.05477315932512283, -0.07857745885848999, -0.001052736770361662, -0.010892337188124657, -0.02345145493745804, 0.096669040620327, 0.0860883891582489, -0.13894593715667725, 0.09370899945497513, 0.08964486420154572, 0.07878966629505157, 0.08944365382194519, -0.019288908690214157, -0.07786829769611359, -0.043186254799366, 0.033137641847133636, 0.01901891455054283, 0.12420984357595444, -0.039697933942079544, 0.03966275230050087, 0.040965113788843155, -0.02369535155594349, 0.0208594873547554, -0.0779976174235344, 0.03547227382659912, 0.025451522320508957, -0.013419056311249733, 0.0522155836224556, -0.03596028685569763, 0.01768462173640728, 0.08731383085250854, 0.0600498728454113, 0.03784669190645218, 0.019416311755776405, -0.05496416613459587, -0.11417433619499207, 0.1614515781402588, -0.12736648321151733, -0.21757090091705322, -0.13575896620750427, 0.006219001021236181, 0.02713099494576454, -0.01611080765724182, 0.005273580551147461, -0.06215012073516846, -0.11567096412181854, -0.08801495283842087, 0.01369257178157568, 0.04681573063135147, -0.08296521753072739, -0.057087384164333344, 0.04934047535061836, 0.039886169135570526, -0.1423034518957138, 0.0192253515124321, 0.04457421973347664, -0.09439462423324585, -0.009149068966507912, 0.07709269970655441, 0.07575637102127075, 0.18514493107795715, 0.023261457681655884, -0.015740202739834785, 0.03361797705292702, 0.2173980325460434, -0.13463890552520752, 0.11001123487949371, 0.1351759135723114, -0.08579442650079727, 0.07642313092947006, 0.2039504498243332, 0.03976231813430786, -0.09841151535511017, 0.03274962678551674, 0.02716667577624321, -0.02665826492011547, -0.23611822724342346, -0.06864923983812332, -0.0025264304131269455, -0.061071690171957016, 0.07785852998495102, 0.09445744007825851, 0.08042383193969727, 0.012628018856048584, -0.09212273359298706, -0.08962614834308624, 0.06138162314891815, 0.1075759083032608, 0.020866168662905693, -0.010555485263466835, 0.08991485834121704, -0.036821044981479645, 0.017841218039393425, 0.0845731794834137, 0.0003158582840114832, 0.16539588570594788, 0.052359748631715775, 0.18166592717170715, 0.08556290715932846, 0.06773394346237183, 0.006052517332136631, 0.011089131236076355, 0.01753428392112255, 0.0396244116127491, -0.00489289965480566, -0.08247264474630356, -0.026122495532035828, 0.11060881614685059, 0.06921198219060898, 0.013344032689929008, 0.0036440896801650524, -0.04114439710974693, 0.08063098043203354, 0.1851523220539093, -0.0015326890861615539, -0.18181483447551727, -0.05811255797743797, 0.07082486152648926, -0.0963631346821785, -0.09890899807214737, -0.007824650965631008, 0.015081853605806828, -0.1664934903383255, 0.03131743520498276, -0.023142162710428238, 0.10727789998054504, -0.1339508295059204, -0.017811620607972145, 0.08476732671260834, 0.07694979757070541, 0.003366870805621147, 0.053349703550338745, -0.17976051568984985, 0.09738051891326904, 0.011453966610133648, 0.06580482423305511, -0.09831780195236206, 0.09574286639690399, -0.00848438311368227, -0.02974962629377842, 0.14086779952049255, -0.0021060993894934654, -0.08418500423431396, -0.07097478210926056, -0.08546790480613708, -0.011803794652223587, 0.13191695511341095, -0.13523392379283905, 0.09175737202167511, -0.038317784667015076, -0.0399005301296711, -0.006670610513538122, -0.08286623656749725, -0.10988686233758926, -0.18138302862644196, 0.06428594887256622, -0.13618691265583038, 0.03443993255496025, -0.10699591785669327, -0.028086431324481964, -0.028450101613998413, 0.185893252491951, -0.24285589158535004, -0.0720391720533371, -0.14559976756572723, -0.09523109346628189, 0.13158145546913147, -0.05046520382165909, 0.08948038518428802, -0.014366953633725643, 0.15924744307994843, 0.021301938220858574, -0.02642022632062435, 0.09679772704839706, -0.08797615766525269, -0.19480395317077637, -0.07075656950473785, 0.15793077647686005, 0.12338215112686157, 0.03408656269311905, -0.0028886415529996157, 0.03741986304521561, -0.017397474497556686, -0.11732777953147888, 0.02374573051929474, 0.1611156016588211, 0.05927205830812454, 0.013477201573550701, -0.027798181399703026, -0.10205331444740295, -0.07143357396125793, -0.027033917605876923, 0.034523505717515945, 0.17596636712551117, -0.07159668952226639, 0.1756334751844406, 0.1423782855272293, -0.05789657682180405, -0.21378444135189056, 0.002020598389208317, 0.026645610108971596, -0.004715518094599247, 0.013092443346977234, -0.19608084857463837, 0.08539512008428574, -0.0034772022627294064, -0.05243929103016853, 0.11376505345106125, -0.17101463675498962, -0.13795442879199982, 0.0834243893623352, 0.04512018337845802, -0.18096978962421417, -0.13552488386631012, -0.09247418493032455, -0.04013165086507797, -0.1711416393518448, 0.09273304045200348, 0.027995865792036057, 0.0148277897387743, 0.030328257009387016, 0.01938215270638466, 0.021149810403585434, -0.04791843518614769, 0.17555223405361176, -0.019769979640841484, 0.020409544929862022, -0.0936831459403038, -0.07518516480922699, 0.024298660457134247, -0.052050165832042694, 0.07289647310972214, -0.011166165582835674, 0.009716099128127098, -0.10240643471479416, -0.03636574000120163, -0.04356086999177933, 0.015470609068870544, -0.09905023872852325, -0.08548939228057861, -0.04109233617782593, 0.09759332984685898, 0.09541726857423782, -0.027944352477788925, -0.02511867880821228, -0.07781163603067398, 0.05217910185456276, 0.21110229194164276, 0.1823381781578064, 0.037121910601854324, -0.06694791465997696, -0.0031468700617551804, -0.015021306462585926, 0.041322823613882065, -0.19275815784931183, 0.061164651066064835, 0.05642986670136452, 0.02123778872191906, 0.10593771189451218, -0.014023703522980213, -0.15881122648715973, -0.07744991779327393, 0.06884031742811203, -0.0642467588186264, -0.196654811501503, 0.005015682429075241, 0.05778492987155914, -0.17450599372386932, -0.041658248752355576, 0.048427291214466095, -0.0025725234299898148, -0.03752648085355759, 0.024587728083133698, 0.09566520899534225, 0.002649798523634672, 0.07892350852489471, 0.06856232136487961, 0.08109209686517715, -0.10347707569599152, 0.08565342426300049, 0.09468026459217072, -0.07799868285655975, 0.02685210108757019, 0.10873907059431076, -0.060533083975315094, -0.039106808602809906, 0.026125194504857063, 0.08475253731012344, 0.01885966956615448, -0.037850234657526016, 0.009164165705442429, -0.10049069672822952, 0.06755121797323227, 0.09617457538843155, 0.03131009265780449, 0.018866542726755142, 0.041336413472890854, 0.04965097829699516, -0.07615652680397034, 0.1230916827917099, 0.027791865170001984, 0.01464198250323534, -0.039888933300971985, -0.03825610503554344, 0.00910575408488512, -0.024327591061592102, -0.006081572268158197, -0.025255419313907623, -0.08725221455097198, -0.01623663492500782, -0.13651569187641144, -0.017142390832304955, -0.059542421251535416, 0.011991783045232296, 0.02953115478157997, -0.030627675354480743, 0.0029965939465910196, 0.007852766662836075, -0.07453011721372604, -0.07039297372102737, -0.016652902588248253, 0.09216049313545227, -0.15959587693214417, 0.022605231031775475, 0.08064928650856018, -0.11696389317512512, 0.09602848440408707, 0.014388787560164928, -0.004279856104403734, 0.023146308958530426, -0.1361229121685028, 0.030282974243164062, -0.039133716374635696, 0.0076028392650187016, 0.03160247579216957, -0.20821495354175568, -0.00022072528372518718, -0.03575378656387329, -0.07267094403505325, -0.008579190820455551, -0.027544893324375153, -0.11273812502622604, 0.10654313117265701, -0.00020246249914634973, -0.07997029274702072, -0.03441299498081207, 0.03191130608320236, 0.0780298188328743, -0.021122561767697334, 0.1514662653207779, -0.012990621849894524, 0.07141751050949097, -0.15981611609458923, -0.012005859054625034, -0.009443454444408417, 0.01767248474061489, -0.02847808226943016, -0.008984408341348171, 0.05071726813912392, -0.018409498035907745, 0.17420686781406403, -0.03575613722205162, 0.020326193422079086, 0.07062411308288574, 0.04051674157381058, -0.03180088475346565, 0.09997711330652237, 0.0412830226123333, 0.019292866811156273, 0.01314468216150999, 0.014155887998640537, -0.03685736656188965, -0.03277828171849251, -0.1917176991701126, 0.07699165493249893, 0.17488610744476318, 0.09975257515907288, -0.015583284199237823, 0.07288500666618347, -0.10419046878814697, -0.09661266952753067, 0.14995692670345306, -0.03819677606225014, -0.0031274498905986547, -0.07430916279554367, 0.12473085522651672, 0.14652439951896667, -0.17896796762943268, 0.06796629726886749, -0.07051600515842438, -0.04126399755477905, -0.11351794749498367, -0.19178161025047302, -0.05950431525707245, -0.05095096677541733, -0.019952403381466866, -0.044757187366485596, 0.07019958645105362, 0.054889533668756485, 0.004235032480210066, -0.006166468840092421, 0.06790349632501602, -0.038526758551597595, -0.005415244027972221, 0.027267970144748688, 0.058661557734012604, 0.006377870216965675, -0.03443073108792305, 0.016949167475104332, -0.00934294331818819, 0.05496647581458092, 0.07217129319906235, 0.04750750586390495, -0.031855326145887375, 0.020676666870713234, -0.040062978863716125, -0.10771127045154572, 0.04477071017026901, -0.02726154401898384, -0.07553227245807648, 0.15417413413524628, 0.01851602829992771, 0.0036449716426432133, -0.01296702865511179, 0.23345546424388885, -0.06655333936214447, -0.09923107177019119, -0.14767496287822723, 0.08246280997991562, -0.03520327806472778, 0.05279837176203728, 0.03916545957326889, -0.10658429563045502, 0.022373082116246223, 0.14715421199798584, 0.16172845661640167, -0.04128354787826538, 0.021392131224274635, 0.03864733874797821, 0.007434586063027382, -0.026255831122398376, 0.043630488216876984, 0.06763601303100586, 0.1565927118062973, -0.04588843509554863, 0.08520833402872086, -0.002393540693446994, -0.09033550322055817, -0.03680314123630524, 0.1124195009469986, -0.006069475784897804, 0.01780354045331478, -0.05511290207505226, 0.11760566383600235, -0.0710512325167656, -0.22188463807106018, 0.04365820810198784, -0.06866282969713211, -0.13262608647346497, -0.026861686259508133, 0.08169858157634735, -0.010405568405985832, 0.02538945898413658, 0.07695822417736053, -0.06821570545434952, 0.1983860284090042, 0.04018443077802658, -0.05937066674232483, -0.05603239685297012, 0.07778695970773697, -0.08396721631288528, 0.2856498956680298, 0.013873557560145855, 0.03625119850039482, 0.11069942265748978, -0.010426776483654976, -0.14402715861797333, 0.01899195834994316, 0.0949523076415062, -0.10360365360975266, 0.04933980107307434, 0.1883028745651245, -0.000888897804543376, 0.1319187730550766, 0.07365170121192932, -0.08785652369260788, 0.0466892346739769, -0.08022354543209076, -0.06671570241451263, -0.09508565068244934, 0.10285025835037231, -0.08357170969247818, 0.14530859887599945, 0.13243870437145233, -0.05487304925918579, 0.010461969301104546, -0.03746109455823898, 0.0432126522064209, -0.0031015127897262573, 0.11444036662578583, 0.006301205139607191, -0.18763801455497742, 0.02736278437077999, -0.02544599212706089, 0.10289593040943146, -0.16350138187408447, -0.09151041507720947, 0.04605849087238312, 0.00767862843349576, -0.07400921732187271, 0.1278243213891983, 0.057374801486730576, 0.033772069960832596, -0.04755678027868271, -0.022046254947781563, -0.008377399295568466, 0.1401062309741974, -0.11112459003925323, -0.006942050997167826 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"license": "apache-2.0", "library_name": "transformers", "datasets": ["jondurbin/truthy-dpo-v0.1"]}
text-generation
vicgalle/zephyr-7b-truthy
[ "transformers", "safetensors", "mistral", "text-generation", "conversational", "dataset:jondurbin/truthy-dpo-v0.1", "arxiv:1910.09700", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T13:48:48+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #mistral #text-generation #conversational #dataset-jondurbin/truthy-dpo-v0.1 #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #mistral #text-generation #conversational #dataset-jondurbin/truthy-dpo-v0.1 #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 84, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #mistral #text-generation #conversational #dataset-jondurbin/truthy-dpo-v0.1 #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]" ]
[ -0.07881546765565872, 0.19828130304813385, -0.003788086585700512, 0.02594563364982605, 0.10099275410175323, -0.004296048078685999, 0.0479087196290493, 0.1277962476015091, 0.00588206248357892, 0.13631753623485565, 0.035890862345695496, 0.15719355642795563, 0.10844240337610245, 0.1793917864561081, 0.007658287417143583, -0.15868471562862396, 0.0245195385068655, -0.10021348297595978, 0.024458106607198715, 0.11435500532388687, 0.12693296372890472, -0.10097949206829071, 0.06654553860425949, -0.04340309277176857, 0.005332469940185547, -0.05228358507156372, -0.07581142336130142, -0.03051694668829441, 0.04182282090187073, 0.029649484902620316, 0.05053135007619858, -0.004885184578597546, 0.07960651814937592, -0.2893868088722229, 0.019171148538589478, 0.05957712233066559, -0.0068930331617593765, 0.081367626786232, 0.08080241084098816, -0.04768414795398712, 0.052873868495225906, -0.09125132113695145, 0.11165368556976318, 0.1035337820649147, -0.07134167850017548, -0.15685921907424927, -0.0693986713886261, 0.10982432961463928, 0.17701099812984467, 0.061307866126298904, -0.03139302507042885, 0.1258116364479065, -0.029040049761533737, 0.03488624840974808, 0.042504508048295975, -0.06233052536845207, -0.054855868220329285, 0.05151001363992691, 0.12500938773155212, 0.05076473951339722, -0.12248334288597107, 0.0024265197571367025, 0.028647491708397865, 0.027724992483854294, 0.09080187231302261, 0.014314192347228527, 0.15132085978984833, 0.03204616904258728, -0.13304442167282104, -0.04573747143149376, 0.06554054468870163, 0.019732391461730003, -0.03674599528312683, -0.24686864018440247, -0.011515004560351372, -0.03573618829250336, -0.03669242933392525, -0.07523033767938614, 0.04231909289956093, 0.010482934303581715, 0.10053340345621109, -0.03943395987153053, -0.079249806702137, -0.020038429647684097, 0.07347044348716736, 0.058268915861845016, 0.015719229355454445, -0.032701823860406876, 0.007287043146789074, 0.10275758057832718, 0.07646103203296661, -0.13315226137638092, -0.047502677887678146, -0.05886125564575195, -0.0792907103896141, -0.03181007131934166, 0.049969371408224106, 0.058802664279937744, 0.056187473237514496, 0.23812977969646454, 0.00020414328901097178, 0.04227973148226738, 0.017442164942622185, -0.0027322338428348303, 0.05583572015166283, 0.0890578180551529, -0.06800514459609985, -0.16957883536815643, -0.018976964056491852, 0.0959894210100174, -0.0005425198469310999, -0.038048259913921356, -0.05217256024479866, 0.052499543875455856, 0.06655170768499374, 0.11139781773090363, 0.15648657083511353, 0.004293535836040974, -0.07813064008951187, -0.07859116047620773, 0.2078639417886734, -0.14715109765529633, 0.034201689064502716, 0.00617571035400033, -0.015117393806576729, -0.04204301908612251, 0.010914871469140053, 0.025545824319124222, -0.043178070336580276, 0.061957795172929764, -0.06763803958892822, -0.045910272747278214, -0.10450239479541779, -0.014932607300579548, 0.04488568380475044, -0.01481498870998621, -0.03474658355116844, -0.04571165889501572, -0.08340682089328766, -0.084851935505867, 0.09068591892719269, -0.06649108231067657, -0.05018448084592819, -0.03816819190979004, -0.08369343727827072, 0.030897997319698334, 0.004291737452149391, 0.05552113056182861, -0.024042882025241852, 0.0485309474170208, -0.023935697972774506, 0.043150316923856735, 0.09938269108533859, 0.039635591208934784, -0.06291232258081436, 0.06383609771728516, -0.17414671182632446, 0.0871361494064331, -0.07548221200704575, 0.038742970675230026, -0.16294081509113312, -0.008647636510431767, 0.050708867609500885, 0.01890251226723194, 0.016494527459144592, 0.15003183484077454, -0.1978839486837387, -0.011560849845409393, 0.16447751224040985, -0.08672865480184555, -0.1205868273973465, 0.038608577102422714, -0.04632066935300827, 0.17060771584510803, 0.04823600500822067, -0.026249073445796967, 0.08270324766635895, -0.15326376259326935, -0.07186512649059296, -0.04507352411746979, -0.017185712233185768, 0.11828253418207169, 0.07537685334682465, -0.06942427903413773, 0.07968953251838684, 0.027863051742315292, -0.050339024513959885, -0.04515686258673668, -0.027160923928022385, -0.10464459657669067, 0.02292945794761181, -0.0703677237033844, 0.033380892127752304, -0.005213726311922073, -0.07840840518474579, 0.004490640480071306, -0.17269867658615112, -0.014996840618550777, 0.08547213673591614, 0.010901116766035557, -0.012243608944118023, -0.100921630859375, 0.013962926343083382, -0.010242554359138012, 0.0026573913637548685, -0.12942355871200562, -0.05195658653974533, 0.03399115055799484, -0.13868989050388336, 0.015431039966642857, -0.1263042688369751, 0.05113783851265907, 0.01730240322649479, -0.04055362194776535, -0.03885255753993988, 0.012320130132138729, 0.013627488166093826, -0.03197228163480759, -0.23036736249923706, -0.03626461327075958, -0.05209866166114807, 0.12405294179916382, -0.17210176587104797, 0.04660525172948837, 0.03528955951333046, 0.154791459441185, -0.008823263458907604, -0.062012650072574615, 0.023054346442222595, -0.06283076107501984, -0.028296422213315964, -0.0581851452589035, 0.016212422400712967, -0.03035099245607853, -0.036525145173072815, 0.029630037024617195, -0.1738484650850296, -0.07187142968177795, 0.10492879152297974, 0.037480615079402924, -0.11284895241260529, -0.04220625385642052, -0.017289940267801285, -0.0842842236161232, -0.027866095304489136, -0.09179391711950302, 0.11602558940649033, 0.064602330327034, 0.025295423343777657, -0.05807238444685936, -0.0957917794585228, 0.01143091544508934, 0.005810038652271032, -0.024226972833275795, 0.07300575822591782, 0.03304094448685646, -0.1464795023202896, 0.10047832131385803, 0.08730784058570862, 0.07242788374423981, 0.10477128624916077, -0.010574106127023697, -0.08535327762365341, -0.054289668798446655, 0.053848814219236374, 0.028533045202493668, 0.12984049320220947, -0.08748336136341095, 0.026734469458460808, 0.03565197438001633, -0.046436019241809845, 0.04171659052371979, -0.05022921413183212, 0.02891300618648529, 0.002196888206526637, -0.0008687818772159517, 0.06161680445075035, -0.047712672501802444, 0.00358104077167809, 0.053274258971214294, 0.07516757398843765, 0.03713718429207802, 0.03479282930493355, -0.049143511801958084, -0.11213786154985428, 0.14144432544708252, -0.11473142355680466, -0.22945551574230194, -0.15714401006698608, 0.012630180455744267, 0.027595121413469315, -0.012676641345024109, 0.0002522561408113688, -0.0617019459605217, -0.09118322283029556, -0.06718011945486069, 0.028629612177610397, 0.030481429770588875, -0.07362861186265945, -0.05192512273788452, 0.0673287957906723, 0.03888237103819847, -0.13425830006599426, 0.02333889715373516, 0.05814511328935623, -0.0503443107008934, -0.01452688779681921, 0.09098145365715027, 0.10071311891078949, 0.15509368479251862, 0.026995470747351646, -0.024426763877272606, 0.03903128206729889, 0.1890534609556198, -0.13888412714004517, 0.10429771989583969, 0.14118772745132446, -0.06478054076433182, 0.08099202811717987, 0.21397748589515686, 0.028014136478304863, -0.08075498789548874, 0.02717246115207672, 0.03598237410187721, -0.0168061014264822, -0.2405443638563156, -0.07052956521511078, -0.009542427957057953, -0.06399513781070709, 0.07753880321979523, 0.06916216015815735, 0.10138556361198425, 0.020266665145754814, -0.0951237678527832, -0.06340327113866806, 0.04275375232100487, 0.11852041631937027, -0.015316975302994251, -0.0040738522075116634, 0.08654475957155228, 0.000752968539018184, 0.009894811548292637, 0.09171147644519806, 0.005475899670273066, 0.15728604793548584, 0.029750864952802658, 0.18468399345874786, 0.08047284930944443, 0.08130007237195969, -0.024692878127098083, 0.022424111142754555, 0.034270837903022766, 0.04760831594467163, 0.0010454743169248104, -0.09034740924835205, -0.006805452983826399, 0.1309634894132614, 0.02666550688445568, 0.00850677490234375, 0.027163676917552948, -0.028908856213092804, 0.06819561868906021, 0.17373552918434143, -0.020335545763373375, -0.18999046087265015, -0.07895176857709885, 0.0819571241736412, -0.07079801708459854, -0.12973712384700775, -0.015118171460926533, 0.019867567345499992, -0.15224464237689972, 0.012756898067891598, -0.047877758741378784, 0.10251205414533615, -0.11504417657852173, -0.011451746337115765, 0.0647401288151741, 0.0570414662361145, 0.01916520670056343, 0.05474560707807541, -0.149484783411026, 0.0977819636464119, 0.037385232746601105, 0.07960565388202667, -0.09108329564332962, 0.09984303265810013, 0.00465016532689333, -0.05566259101033211, 0.174989253282547, -0.006504591554403305, -0.03550473228096962, -0.09755636751651764, -0.12537525594234467, -0.029732022434473038, 0.11871903389692307, -0.1574956625699997, 0.09242075681686401, -0.030641116201877594, -0.035121552646160126, 0.0009906096383929253, -0.07698814570903778, -0.1236131489276886, -0.1708577573299408, 0.07721659541130066, -0.12350669503211975, 0.048636823892593384, -0.09172387421131134, -0.05109061300754547, -0.003261775244027376, 0.2252921313047409, -0.23203054070472717, -0.08419772982597351, -0.1426885426044464, -0.07339078933000565, 0.15236058831214905, -0.06525767594575882, 0.08257818967103958, -0.0021147867664694786, 0.13715380430221558, 0.001774284290149808, -0.025268927216529846, 0.08819132298231125, -0.089314766228199, -0.17512951791286469, -0.0634731575846672, 0.14017701148986816, 0.12897665798664093, 0.021026894450187683, -0.008259249851107597, 0.028918003663420677, -0.025781404227018356, -0.11042740195989609, 0.024032382294535637, 0.18602684140205383, 0.08706289529800415, -0.003113867249339819, -0.026645395904779434, -0.15675631165504456, -0.07985081523656845, -0.042571332305669785, -0.00569736585021019, 0.1937917321920395, -0.06679373979568481, 0.16908712685108185, 0.1621122658252716, -0.06503459811210632, -0.2172807902097702, -0.0216207355260849, 0.034951429814100266, -0.009944216348230839, 0.03756925091147423, -0.17392128705978394, 0.09013393521308899, -0.04237564280629158, -0.06530312448740005, 0.11577162891626358, -0.1340962052345276, -0.14075429737567902, 0.07945678383111954, 0.03325343504548073, -0.17837609350681305, -0.13257193565368652, -0.11716437339782715, -0.013849650509655476, -0.12169323116540909, 0.08157778531312943, -0.01762000471353531, 0.00006085873246775009, 0.024526460096240044, 0.005087598692625761, 0.039726510643959045, -0.05819311738014221, 0.17777453362941742, -0.023438790813088417, 0.00033240654738619924, -0.07579733431339264, -0.09002464264631271, 0.06345563381910324, -0.05954119563102722, 0.09047460556030273, -0.01610659249126911, 0.021563906222581863, -0.10401412099599838, -0.04758157953619957, -0.05587709695100784, 0.026472527533769608, -0.09847619384527206, -0.10183899849653244, -0.019374975934624672, 0.09902661293745041, 0.10612063854932785, -0.017491566017270088, 0.037717778235673904, -0.058098189532756805, 0.05849234759807587, 0.2573933005332947, 0.18561600148677826, 0.06237258389592171, -0.04560523480176926, -0.004375700373202562, -0.02589179389178753, 0.045906614512205124, -0.1884150356054306, 0.050544463098049164, 0.05442243441939354, 0.019697144627571106, 0.07854799181222916, -0.01410416979342699, -0.154921293258667, -0.061627428978681564, 0.07190555334091187, -0.05120985954999924, -0.18615977466106415, -0.026090208441019058, 0.037609562277793884, -0.1925758421421051, -0.019302330911159515, 0.07155706733465195, 0.009145748801529408, -0.037576209753751755, 0.0196884423494339, 0.09776037186384201, -0.0009301623213104904, 0.08760736137628555, 0.05622025579214096, 0.08715860545635223, -0.08696423470973969, 0.06254485994577408, 0.10142189264297485, -0.05448020249605179, 0.04566512629389763, 0.1121007651090622, -0.049286868423223495, -0.04838484153151512, 0.06194585934281349, 0.05679060146212578, 0.019992122426629066, -0.031451266258955, 0.0070560690946877, -0.011169198900461197, 0.04698837175965309, 0.08127573132514954, 0.01714867353439331, 0.007729835342615843, 0.06085601821541786, 0.045793913304805756, -0.05725101754069328, 0.12884485721588135, 0.04548371955752373, 0.015413968823850155, -0.05959885194897652, -0.036754339933395386, 0.0016637861263006926, -0.012131733819842339, -0.02155674621462822, -0.011944356374442577, -0.08570700883865356, -0.013417158275842667, -0.15673376619815826, 0.030557289719581604, -0.10762830078601837, -0.00041262651211582124, 0.006363848224282265, -0.030188586562871933, 0.009064258076250553, 0.01401964109390974, -0.05151880159974098, -0.08674784004688263, -0.015339035540819168, 0.11590209603309631, -0.16002780199050903, 0.014311563223600388, 0.07994300872087479, -0.09922817349433899, 0.08586767315864563, -0.012157579883933067, 0.008687136694788933, 0.010018114000558853, -0.14188933372497559, 0.05494757741689682, -0.040540438145399094, 0.00593084841966629, 0.008361957967281342, -0.1925685554742813, -0.016736449673771858, -0.03982504829764366, -0.05960344895720482, -0.002206806093454361, 0.004871213808655739, -0.11585628986358643, 0.06224488839507103, 0.011133942753076553, -0.04489300027489662, -0.0331563763320446, 0.03556939959526062, 0.09809030592441559, -0.03379448875784874, 0.09875980764627457, -0.007952249608933926, 0.07526570558547974, -0.17710962891578674, -0.003297174582257867, -0.013335742056369781, 0.04430251941084862, -0.012741725891828537, -0.02896476909518242, 0.04576569050550461, -0.024412399157881737, 0.14052896201610565, -0.04272871091961861, 0.0731034129858017, 0.04841770604252815, 0.00845903530716896, -0.0006744765560142696, 0.08067823201417923, 0.04936865717172623, 0.010209578089416027, 0.010102658532559872, 0.009001818485558033, -0.0203599501401186, -0.05675429105758667, -0.16490603983402252, 0.016566429287195206, 0.1838306337594986, 0.09017551690340042, 0.0029047399293631315, 0.06561041623353958, -0.12029856443405151, -0.11793289333581924, 0.13760054111480713, -0.03872014954686165, -0.03176084905862808, -0.08670827746391296, 0.12873926758766174, 0.1325829029083252, -0.17378957569599152, 0.08253435790538788, -0.05428522825241089, -0.041164152324199677, -0.0871848538517952, -0.20769929885864258, -0.05384355038404465, -0.03117726370692253, -0.014087420888245106, -0.05315567925572395, 0.05790875479578972, 0.0551043339073658, -0.020521970465779305, -0.013969428837299347, 0.08262757956981659, 0.01174824871122837, -0.02660449780523777, 0.04565690830349922, 0.056925710290670395, 0.013070810586214066, -0.0644959956407547, 0.025732891634106636, -0.0058279032818973064, 0.059858012944459915, 0.07507255673408508, 0.02485421858727932, -0.03682160750031471, 0.017871694639325142, -0.028778541833162308, -0.12587258219718933, 0.04013444110751152, -0.012291443534195423, -0.057522859424352646, 0.1866769939661026, 0.027421627193689346, -0.0016996754566207528, -0.01473489310592413, 0.2264416217803955, -0.07107027620077133, -0.08879691362380981, -0.13616283237934113, 0.060508206486701965, -0.04538299888372421, 0.034724973142147064, 0.028873175382614136, -0.11759581416845322, 0.00371868209913373, 0.15012221038341522, 0.1376749873161316, -0.010608833283185959, 0.011741950176656246, 0.042162857949733734, 0.007835287600755692, -0.04732910916209221, 0.024009786546230316, 0.043765999376773834, 0.1749413162469864, -0.06099776551127434, 0.07447808980941772, 0.019976118579506874, -0.10050734132528305, -0.014391222968697548, 0.07854747027158737, -0.03487072139978409, 0.0376862958073616, -0.05841081589460373, 0.11774079501628876, -0.06283203512430191, -0.23269148170948029, 0.038614049553871155, -0.09184236079454422, -0.12224751710891724, -0.036600515246391296, 0.06463715434074402, -0.026931094005703926, 0.016587495803833008, 0.09203383326530457, -0.03750869631767273, 0.1967022716999054, 0.02298274077475071, -0.07267346233129501, -0.03440467268228531, 0.048213981091976166, -0.14733335375785828, 0.2720259726047516, 0.001460630097426474, 0.04151538386940956, 0.118688203394413, -0.017858754843473434, -0.1480560600757599, -0.006574925500899553, 0.09749066084623337, -0.08423356711864471, 0.06818953901529312, 0.2088891863822937, -0.010896257124841213, 0.11990156024694443, 0.07482453435659409, -0.055415112525224686, 0.024019334465265274, -0.05523483827710152, -0.07215474545955658, -0.11175215989351273, 0.08484652638435364, -0.0837579220533371, 0.16375470161437988, 0.10128376632928848, -0.06383172422647476, 0.0015340227400884032, -0.033121850341558456, 0.06503324210643768, -0.008601132780313492, 0.12789203226566315, -0.008612642996013165, -0.19350647926330566, 0.032298408448696136, 0.037629734724760056, 0.11666148900985718, -0.18992139399051666, -0.07883117347955704, 0.07059340178966522, -0.004890486132353544, -0.07717996835708618, 0.11404167860746384, 0.06026830896735191, 0.01665038801729679, -0.03991955891251564, -0.08306323736906052, -0.018458427861332893, 0.12130540609359741, -0.1141250804066658, -0.024323677644133568 ]
null
null
stable-baselines3
# **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
{"library_name": "stable-baselines3", "tags": ["LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "PPO", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "LunarLander-v2", "type": "LunarLander-v2"}, "metrics": [{"type": "mean_reward", "value": "257.25 +/- 21.93", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
Benfnash/ppo-LunarLander-v2
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
2024-02-10T13:50:29+00:00
[]
[]
TAGS #stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
# PPO Agent playing LunarLander-v2 This is a trained model of a PPO agent playing LunarLander-v2 using the stable-baselines3 library. ## Usage (with Stable-baselines3) TODO: Add your code
[ "# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.", "## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ "TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n", "# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.", "## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ 39, 41, 17 ]
[ "passage: TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ 0.03942384943366051, 0.04900386184453964, -0.005304091144353151, 0.026427261531352997, 0.107408307492733, -0.026511888951063156, 0.11188238859176636, 0.0814051404595375, 0.10722193866968155, 0.04762078449130058, 0.08338645845651627, 0.06030960753560066, 0.05080918222665787, 0.2571701407432556, 0.04754156619310379, -0.22987541556358337, 0.036159250885248184, -0.04869936779141426, 0.12395193427801132, 0.07178173214197159, -0.0038484656251966953, -0.06485428661108017, 0.020415637642145157, -0.013290755450725555, 0.05367108806967735, 0.04282612353563309, -0.01716216839849949, -0.08207534998655319, 0.07169748842716217, -0.06345846503973007, 0.06986866891384125, 0.07677983492612839, 0.13218913972377777, -0.17832116782665253, 0.029566360637545586, 0.02571309357881546, -0.07189024239778519, 0.01342033501714468, 0.008019951172173023, 0.05120139941573143, 0.17303818464279175, 0.019879888743162155, 0.07844575494527817, -0.0025605305563658476, -0.15412317216396332, -0.018950799480080605, 0.0436202734708786, 0.12546207010746002, 0.08808347582817078, 0.04605821147561073, 0.01970590092241764, 0.17503218352794647, -0.054352790117263794, -0.028833400458097458, 0.21759237349033356, -0.2881564497947693, -0.031460098922252655, 0.321048766374588, 0.06997483223676682, 0.09725230932235718, -0.07540661096572876, -0.03619609400629997, 0.007783263456076384, -0.013137873262166977, -0.028666524216532707, -0.07447073608636856, 0.17313385009765625, 0.05152064561843872, -0.05057951435446739, -0.09541505575180054, 0.16948209702968597, 0.006921638268977404, 0.0018855923553928733, -0.019282981753349304, 0.009060598909854889, 0.07402525842189789, -0.016097044572234154, -0.07255112379789352, 0.057438433170318604, 0.05330665782094002, 0.019649166613817215, -0.1435653269290924, -0.10762494057416916, -0.022740179672837257, -0.008012006990611553, 0.17786912620067596, -0.009255532175302505, 0.042902372777462006, 0.003065188182517886, 0.10384012013673782, -0.12480384111404419, -0.03354184702038765, -0.0454259067773819, -0.07565800100564957, -0.0223417766392231, -0.02058211714029312, -0.03580251708626747, 0.07184842973947525, 0.11971849203109741, 0.027368178591132164, 0.09350208193063736, 0.047715865075588226, -0.03206788748502731, 0.06343851238489151, 0.05555703118443489, 0.14222665131092072, 0.05807621404528618, 0.012854371219873428, 0.13179877400398254, 0.055213116109371185, 0.033023182302713394, -0.0613492950797081, -0.18252409994602203, 0.07489913702011108, -0.07031869143247604, 0.007941240444779396, 0.12051256000995636, -0.04480670019984245, -0.1183447614312172, -0.037500523030757904, -0.017392054200172424, -0.06224250793457031, -0.025395862758159637, 0.0547584593296051, -0.02883218228816986, -0.03973718360066414, 0.0011496668448671699, 0.09384800493717194, 0.00953749567270279, -0.1752052903175354, 0.03303423151373863, -0.025042934343218803, -0.10782608389854431, 0.009975161403417587, 0.0022444494534283876, 0.03394931182265282, 0.04408763721585274, -0.11822668462991714, -0.30899152159690857, -0.07652641832828522, 0.05490870401263237, -0.06516939401626587, -0.18425025045871735, -0.13193942606449127, 0.02454492449760437, -0.09037084132432938, -0.044885024428367615, -0.12759265303611755, -0.028549788519740105, 0.01743689924478531, 0.011519349180161953, 0.10758619755506516, -0.0106219332665205, -0.012188062071800232, -0.1571401208639145, 0.008273907005786896, -0.20951123535633087, 0.0890483483672142, -0.019150104373693466, 0.037884220480918884, -0.032381169497966766, -0.07404014468193054, 0.030707746744155884, 0.052499737590551376, -0.01474119070917368, 0.13510210812091827, -0.15592676401138306, -0.03691192343831062, -0.007996266707777977, -0.13611900806427002, -0.04786273464560509, -0.10358831286430359, -0.04357128217816353, 0.13354332745075226, 0.018664736300706863, 0.15356586873531342, -0.08709818124771118, -0.0722038671374321, 0.20489206910133362, -0.010411538183689117, -0.12820468842983246, -0.076752208173275, 0.10165707021951675, 0.021510310471057892, -0.056606587022542953, -0.02523270808160305, -0.1839766949415207, -0.0152357779443264, -0.04550420492887497, -0.047039128839969635, 0.01796751655638218, -0.010888241231441498, 0.13837894797325134, 0.08494598418474197, 0.05018039792776108, -0.06086122244596481, -0.006730288732796907, 0.10779471695423126, 0.08823856711387634, 0.008680110797286034, 0.023406028747558594, -0.05774238705635071, 0.09552932530641556, -0.04003755748271942, -0.0142367510125041, -0.08283266425132751, -0.036246106028556824, -0.026256313547492027, 0.17507147789001465, 0.09440762549638748, 0.2257927656173706, 0.09567736834287643, 0.039160262793302536, 0.031270865350961685, -0.13181598484516144, -0.1425403207540512, -0.0017254541162401438, 0.09020978957414627, -0.14270411431789398, -0.04119925573468208, -0.08974775671958923, -0.17768175899982452, -0.12202505767345428, 0.0006432619411498308, -0.17960017919540405, 0.06390921026468277, 0.05408334732055664, -0.035177867859601974, 0.03272094577550888, 0.13032332062721252, -0.011533179320394993, -0.03967514634132385, 0.0831870287656784, 0.0379033200442791, -0.041234664618968964, -0.021742934361100197, 0.11885567009449005, 0.15673065185546875, 0.13124459981918335, -0.03511447086930275, 0.004914294462651014, 0.07076404243707657, -0.02309088408946991, 0.06539414077997208, 0.0558244064450264, 0.20973342657089233, 0.188301220536232, 0.038996949791908264, 0.008822928182780743, -0.07048165798187256, 0.0855446457862854, -0.0742373839020729, -0.14302679896354675, -0.05579735338687897, 0.08729292452335358, 0.016605578362941742, 0.023469142615795135, 0.08711627870798111, 0.024545932188630104, 0.09132762253284454, 0.15968108177185059, 0.01990218088030815, -0.09659269452095032, -0.050218869000673294, 0.01175848301500082, 0.027713103219866753, 0.04794301092624664, -0.04514073207974434, -0.00937939714640379, 0.017020760104060173, -0.10303554683923721, 0.031789086759090424, -0.1413339376449585, -0.1358717679977417, 0.044326696544885635, 0.003906996920704842, 0.010907664895057678, 0.02786896750330925, -0.0038291432429105043, 0.019039705395698547, 0.04351753741502762, -0.06975466758012772, 0.047416772693395615, -0.024745507165789604, -0.020031947642564774, 0.03340689837932587, -0.057257164269685745, -0.205775648355484, -0.17696654796600342, 0.00013708483311347663, -0.09910997003316879, 0.10194740444421768, 0.018308809027075768, -0.12373185902833939, 0.047737859189510345, -0.05822649225592613, 0.027574289590120316, -0.01875593699514866, -0.049130141735076904, 0.10507171601057053, 0.1525275856256485, -0.016146350651979446, 0.018018173053860664, -0.04865182936191559, -0.10157987475395203, -0.19632206857204437, 0.0691583976149559, 0.04680244252085686, 0.014610917307436466, 0.10669491440057755, 0.018072687089443207, 0.02367905154824257, -0.007674071006476879, -0.016521066427230835, -0.011659215204417706, -0.08781040459871292, 0.31909599900245667, 0.04510033503174782, -0.025173069909214973, 0.02041010931134224, -0.0043001663871109486, -0.028083480894565582, 0.03263787180185318, -0.0985708013176918, -0.07548979669809341, -0.08774089068174362, -0.04367410019040108, -0.09784720093011856, 0.053299110382795334, 0.05916472524404526, 0.003188040340319276, -0.07727594673633575, 0.04221395403146744, 0.11369874328374863, -0.0923808291554451, -0.07137343287467957, 0.07477962225675583, 0.0972946360707283, -0.07331304252147675, 0.00012658814375754446, 0.00874367356300354, 0.023951783776283264, 0.037102166563272476, 0.06778035312891006, -0.03966575115919113, 0.08589404821395874, -0.19917890429496765, 0.0372927263379097, 0.106058269739151, 0.023754918947815895, 0.0638108178973198, 0.07643651217222214, -0.1058402881026268, -0.008500572293996811, -0.032518330961465836, -0.21341575682163239, 0.1668180525302887, 0.1355515867471695, 0.06788124144077301, -0.025637222453951836, -0.00461410591378808, -0.0649740919470787, 0.05773647129535675, 0.02723747305572033, -0.14758841693401337, 0.004883295856416225, 0.06064270809292793, 0.026899009943008423, 0.01614922471344471, 0.07971042394638062, 0.014697225764393806, -0.1801026314496994, -0.014406266622245312, 0.10730406641960144, 0.002390873385593295, 0.0053148469887673855, -0.03175045922398567, -0.1755964607000351, 0.0751047357916832, 0.004285442177206278, 0.07233936339616776, -0.1676585078239441, 0.14297930896282196, -0.10089799761772156, 0.07726949453353882, -0.004285062663257122, -0.021311495453119278, 0.02507244050502777, -0.0541163794696331, 0.15163759887218475, 0.01058570109307766, -0.021810131147503853, -0.1200498715043068, -0.1717042326927185, -0.019227758049964905, -0.11788936704397202, -0.11679866164922714, 0.050424277782440186, 0.062185097485780716, 0.04923136904835701, -0.061147067695856094, 0.1518532931804657, -0.047422297298908234, 0.060713399201631546, -0.06893875449895859, -0.06755045056343079, 0.03764858841896057, -0.12588608264923096, -0.08176055550575256, 0.05573027580976486, 0.19166934490203857, 0.15833087265491486, -0.02816431224346161, -0.03472423925995827, -0.047419581562280655, -0.006212298292666674, -0.007802055217325687, 0.0275666993111372, 0.023223137483000755, 0.07315318286418915, -0.07681374251842499, -0.11649256944656372, 0.033787861466407776, -0.06713802367448807, -0.055589709430933, -0.015439179725944996, 0.1513158082962036, 0.04671623185276985, 0.07720734924077988, -0.018946662545204163, 0.03887668624520302, -0.001724981120787561, -0.056474871933460236, 0.16197094321250916, 0.03885216265916824, -0.05193585529923439, 0.06837689876556396, 0.053174007683992386, 0.043745119124650955, 0.03011113777756691, -0.026783017441630363, 0.206032395362854, 0.1980147808790207, 0.014206883497536182, 0.2175983190536499, 0.03177616000175476, -0.03772832080721855, -0.1300560086965561, -0.065880686044693, -0.006372632458806038, 0.03559038043022156, 0.08070417493581772, -0.18207235634326935, -0.015011128038167953, -0.05689644813537598, -0.034518610686063766, -0.15059494972229004, -0.28553900122642517, -0.05957856774330139, 0.20075850188732147, 0.14706264436244965, 0.27519428730010986, -0.10432573407888412, 0.035197313874959946, 0.02663275972008705, -0.04912831634283066, -0.006501141935586929, 0.00018665487004909664, 0.10268618166446686, -0.15421873331069946, 0.1176437959074974, 0.08486983180046082, -0.019002694636583328, 0.01058861706405878, -0.1619086116552353, 0.00936629343777895, -0.12191236019134521, 0.05354422330856323, 0.1400289237499237, -0.048128653317689896, -0.054873593151569366, 0.14033560454845428, -0.024562934413552284, -0.22685599327087402, -0.04648222774267197, -0.043600670993328094, -0.010640020482242107, 0.026607351377606392, -0.1013401448726654, 0.04101909324526787, 0.1330099105834961, 0.009380043484270573, 0.1147187277674675, 0.11749245226383209, -0.052566803991794586, 0.10792597383260727, 0.2257719188928604, -0.018785694614052773, 0.04689010605216026, -0.12743118405342102, -0.0012336712097749114, -0.028270328417420387, 0.013657891191542149, -0.09504974633455276, -0.09938385337591171, 0.02366873063147068, 0.02872389927506447, 0.009118586778640747, 0.0921793207526207, -0.029922157526016235, 0.0759170651435852, 0.06817561388015747, -0.13014446198940277, -0.16288450360298157, 0.015828335657715797, -0.007344507612287998, 0.08354310691356659, 0.00027861111448146403, 0.08878035843372345, -0.11932205408811569, -0.018093237653374672, -0.03153328225016594, -0.03319635987281799, -0.130486860871315, -0.07138993591070175, 0.06156524643301964, 0.028095467016100883, -0.06602972000837326, 0.1398407518863678, 0.026440169662237167, 0.15942534804344177, 0.049197953194379807, 0.012499804608523846, 0.07227300107479095, -0.05345509201288223, 0.1283530443906784, 0.13818155229091644, -0.00868943240493536, -0.05460423603653908, -0.1013643890619278, -0.10236792266368866, 0.08925779908895493, -0.05773641914129257, 0.07476430386304855, -0.14885357022285461, -0.06675903499126434, 0.015772046521306038, 0.016141414642333984, -0.09562095999717712, 0.02571965754032135, -0.01625603251159191, -0.18119946122169495, 0.056570518761873245, -0.048285093158483505, 0.0440407395362854, -0.06347788125276566, -0.1110161691904068, -0.17226378619670868, 0.06091433763504028, 0.08593481779098511, -0.053876690566539764, -0.12229149043560028, 0.011023230850696564, -0.00012518465518951416, -0.06341652572154999, -0.05023367330431938, 0.09722746908664703, -0.11020902544260025, 0.031452205032110214, -0.012567701749503613, 0.08853451162576675, -0.03510405123233795, -0.011538895778357983, 0.044220831245183945, -0.08039166033267975, -0.009481523185968399, 0.03534642979502678, -0.026372017338871956, -0.04127239063382149, -0.2689029574394226, 0.0036654395516961813, 0.0341104120016098, 0.02497158572077751, 0.07856601476669312, 0.011906822212040424, 0.021174922585487366, 0.03993808850646019, -0.15396519005298615, -0.013395369984209538, 0.14574195444583893, -0.07689505815505981, -0.022186370566487312, 0.05703273415565491, -0.09054436534643173, 0.013882770203053951, -0.030287226662039757, 0.1345842480659485, 0.023923413828015327, 0.06404478847980499, -0.0851147472858429, 0.10106813907623291, -0.1451139897108078, -0.04998219385743141, -0.01244612317532301, 0.09761348366737366, 0.07019034773111343, -0.10272270441055298, 0.014697125181555748, 0.04210108891129494, 0.19416837394237518, 0.016384804621338844, -0.0356343574821949, -0.03396720811724663, 0.004015897400677204, 0.22076453268527985, 0.03044266067445278, 0.10457023978233337, 0.07281364500522614, -0.026583973318338394, 0.12624378502368927, 0.09929762035608292, 0.11280370503664017, -0.055645186454057693, 0.13904185593128204, 0.04667386785149574, 0.038641396909952164, 0.0614289753139019, 0.06836545467376709, 0.09098632633686066, -0.0008288522367365658, 0.1138714924454689, 0.013811973854899406, -0.02422109805047512, -0.021335409954190254, 0.17759373784065247, 0.10501719266176224, -0.14769648015499115, 0.029047364369034767, -0.01258957851678133, 0.039933037012815475, -0.014194529503583908, -0.15634691715240479, -0.07240267097949982, -0.3315149247646332, 0.1226184144616127, -0.07119352370500565, 0.019930170848965645, 0.007913772016763687, -0.037425633519887924, -0.03296699747443199, -0.04477746784687042, 0.13151589035987854, -0.013641550205647945, -0.006079165264964104, -0.04815853759646416, -0.015360191464424133, -0.11607866734266281, -0.11200575530529022, -0.013207737356424332, -0.13671602308750153, -0.010119039565324783, 0.05595948174595833, 0.003977729007601738, 0.01821410097181797, -0.03142618387937546, 0.0024383175186812878, 0.06541839241981506, -0.05751744285225868, 0.056182678788900375, 0.12097269296646118, 0.08766137808561325, -0.1058853268623352, 0.031048951670527458, 0.2011747509241104, 0.04359564557671547, -0.12483977526426315, 0.01449228823184967, 0.1819491684436798, 0.004885740112513304, 0.017068125307559967, -0.006097703706473112, -0.0540788508951664, -0.07554277032613754, 0.1251034289598465, 0.08296554535627365, -0.09985227137804031, 0.015833314508199692, -0.0726347416639328, -0.01594804972410202, -0.06374675035476685, 0.10130585730075836, 0.09538925439119339, 0.04440245032310486, -0.10621760785579681, -0.08487539738416672, -0.10891728103160858, 0.040588874369859695, -0.08629853278398514, -0.07311757653951645, 0.09629398584365845, -0.07057105004787445, -0.07029950618743896, 0.025521177798509598, -0.17978744208812714, -0.009467960335314274, 0.1711762249469757, -0.24654000997543335, -0.0916430801153183, -0.10857923328876495, 0.14477859437465668, 0.016497576609253883, 0.1013975441455841, -0.006207061931490898, -0.007889035157859325, -0.20577777922153473, 0.024890204891562462, -0.05293011665344238, -0.02073732763528824, 0.07814782857894897, -0.09476397186517715, 0.22629831731319427, -0.08276885002851486, 0.020940175279974937, 0.012659613974392414, 0.0870661810040474, -0.030675338581204414, 0.09283176809549332, -0.03660329803824425, -0.12576518952846527, -0.03620953485369682, 0.03001813031733036, 0.013904244638979435, 0.10071761906147003, 0.09772487729787827, -0.03414725139737129, 0.03389119729399681, 0.09747414290904999, 0.04172342270612717, -0.023843804374337196, 0.0360250361263752, -0.17077107727527618, 0.02182629331946373, -0.018498148769140244, -0.06935930997133255, 0.03687669709324837, -0.06603235751390457, 0.1639697551727295, 0.04022442549467087, 0.0670473501086235, -0.036152735352516174, 0.0073931049555540085, -0.014454689808189869, -0.013775371946394444, -0.026180334389209747, -0.17259705066680908, -0.10422050207853317, -0.1347656100988388, -0.012701659463346004, -0.034971047192811966, 0.04591470584273338, 0.023234914988279343, -0.0003200018545612693, -0.014577031135559082, -0.12090865522623062, 0.04360328987240791, 0.11146783083677292, -0.04631396010518074, -0.026193076744675636 ]
null
null
null
# **Reinforce** Agent playing **Pixelcopter-PLE-v0** This is a trained model of a **Reinforce** agent playing **Pixelcopter-PLE-v0** . To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: https://huggingface.co/deep-rl-course/unit4/introduction
{"tags": ["Pixelcopter-PLE-v0", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class"], "model-index": [{"name": "PixelCopter-v1", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "Pixelcopter-PLE-v0", "type": "Pixelcopter-PLE-v0"}, "metrics": [{"type": "mean_reward", "value": "16.70 +/- 12.17", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
leejeewoo/PixelCopter-v1
[ "Pixelcopter-PLE-v0", "reinforce", "reinforcement-learning", "custom-implementation", "deep-rl-class", "model-index", "region:us" ]
2024-02-10T13:51:01+00:00
[]
[]
TAGS #Pixelcopter-PLE-v0 #reinforce #reinforcement-learning #custom-implementation #deep-rl-class #model-index #region-us
# Reinforce Agent playing Pixelcopter-PLE-v0 This is a trained model of a Reinforce agent playing Pixelcopter-PLE-v0 . To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL
[ "# Reinforce Agent playing Pixelcopter-PLE-v0\n This is a trained model of a Reinforce agent playing Pixelcopter-PLE-v0 .\n To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL" ]
[ "TAGS\n#Pixelcopter-PLE-v0 #reinforce #reinforcement-learning #custom-implementation #deep-rl-class #model-index #region-us \n", "# Reinforce Agent playing Pixelcopter-PLE-v0\n This is a trained model of a Reinforce agent playing Pixelcopter-PLE-v0 .\n To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL" ]
[ 41, 58 ]
[ "passage: TAGS\n#Pixelcopter-PLE-v0 #reinforce #reinforcement-learning #custom-implementation #deep-rl-class #model-index #region-us \n# Reinforce Agent playing Pixelcopter-PLE-v0\n This is a trained model of a Reinforce agent playing Pixelcopter-PLE-v0 .\n To learn to use this model and train yours check Unit 4 of the Deep Reinforcement Learning Course: URL" ]
[ 0.0073175891302526, -0.2259262204170227, -0.0017347558168694377, 0.05054566636681557, 0.0658537745475769, -0.055378563702106476, 0.1412602812051773, 0.05916554853320122, -0.04990595206618309, 0.059261854737997055, 0.14166708290576935, 0.03996060788631439, 0.022112762555480003, 0.1513713151216507, 0.09764605015516281, -0.2469022423028946, 0.07438477873802185, 0.01641594059765339, 0.008152224123477936, 0.09583204984664917, 0.060265738517045975, -0.1405058205127716, 0.037032704800367355, -0.01332044042646885, -0.13650871813297272, 0.0010478810872882605, -0.021802188828587532, -0.03625129908323288, 0.15681709349155426, 0.006844013463705778, 0.09602472931146622, -0.001560068572871387, 0.06475798785686493, -0.12438877671957016, 0.05466329678893089, 0.06455880403518677, -0.06293967366218567, 0.058029334992170334, -0.057374246418476105, 0.11959903687238693, 0.04641333222389221, -0.01578129455447197, 0.054811324924230576, 0.010941818356513977, -0.14131468534469604, -0.006710252724587917, 0.007013716734945774, 0.15098218619823456, 0.1339312642812729, 0.01409265398979187, -0.0014771400019526482, 0.1363491266965866, -0.16774429380893707, 0.045684073120355606, 0.061802688986063004, -0.2633039951324463, -0.04168876260519028, 0.12259352207183838, 0.08951573073863983, 0.06848238408565521, -0.060910262167453766, 0.07636868953704834, 0.049813780933618546, 0.013985024765133858, 0.023094501346349716, -0.042509064078330994, -0.040479615330696106, 0.02289252169430256, -0.0921095609664917, -0.05999262258410454, 0.11517233401536942, -0.006806366611272097, 0.03735918551683426, -0.12476086616516113, -0.015330453403294086, -0.07314357161521912, -0.05917041376233101, -0.082573801279068, 0.07563583552837372, 0.030191516503691673, -0.048283837735652924, -0.08895846456289291, -0.056533291935920715, -0.11489585787057877, -0.023082571104168892, -0.07226225733757019, 0.005096882116049528, -0.03157244250178337, -0.035645097494125366, 0.09446526318788528, -0.0021088174544274807, -0.015028090216219425, -0.03452150896191597, -0.05930153280496597, -0.04213470220565796, -0.02359505370259285, -0.03510070592164993, -0.059062156826257706, 0.054655663669109344, 0.0680202916264534, 0.04938843473792076, 0.09133565425872803, -0.0467856265604496, 0.1667373925447464, -0.03256719931960106, 0.08078566938638687, -0.011897698976099491, 0.2012830525636673, 0.11370102316141129, 0.12129533290863037, 0.06716908514499664, -0.05294690653681755, -0.16726544499397278, 0.039163749665021896, 0.12641896307468414, 0.07664673775434494, -0.032492902129888535, 0.018162984400987625, -0.12440363317728043, 0.05439428985118866, -0.14826108515262604, -0.06745084375143051, 0.024251462891697884, 0.01822635903954506, -0.060682263225317, 0.03656952083110809, -0.0028792342636734247, 0.003339326474815607, 0.004654870834201574, -0.16432709991931915, -0.05568019300699234, 0.028964387252926826, -0.15712425112724304, -0.06656725704669952, 0.06277995556592941, -0.10113482922315598, -0.012132617644965649, -0.16982388496398926, -0.16305199265480042, -0.03628521412611008, 0.017857929691672325, -0.040613796561956406, -0.056917786598205566, -0.14010562002658844, -0.019415250048041344, -0.045320261269807816, -0.004312154371291399, 0.044072363525629044, 0.0020940210670232773, 0.04635847359895706, 0.0066573889926075935, 0.09289347380399704, 0.010714372619986534, -0.0014722738415002823, -0.04595406726002693, 0.0909833237528801, -0.30731555819511414, 0.07525643706321716, -0.08645553886890411, 0.05539081245660782, -0.057316381484270096, -0.0926317572593689, -0.007509906310588121, 0.06277763843536377, 0.060464419424533844, 0.20788121223449707, -0.2800109386444092, -0.07025618106126785, 0.13655538856983185, -0.09533236175775528, -0.13146020472049713, 0.0513952374458313, -0.050213608890771866, 0.07593657076358795, 0.027370907366275787, 0.140700101852417, -0.028026295825839043, -0.15554022789001465, 0.06281048059463501, 0.04586128890514374, -0.11356306821107864, 0.019295670092105865, 0.03597676753997803, 0.06723599135875702, 0.05744141340255737, -0.036986757069826126, -0.04105675220489502, 0.08096802979707718, -0.07076814025640488, -0.037564266473054886, 0.04588831216096878, -0.0579565204679966, 0.1630958467721939, 0.033971156924963, 0.09856503456830978, -0.04149768501520157, -0.07435470074415207, -0.005698562134057283, 0.038746561855077744, -0.08962973952293396, 0.025353478267788887, -0.18320298194885254, 0.2423991560935974, -0.02621818706393242, 0.027546977624297142, -0.16845986247062683, -0.0588528998196125, 0.011087946593761444, 0.21568740904331207, 0.030399197712540627, 0.12989304959774017, 0.07485637813806534, -0.01250512059777975, 0.014156299643218517, -0.06183977797627449, -0.1972363442182541, -0.03247830644249916, 0.008314179256558418, -0.058311350643634796, -0.04934588819742203, -0.0900716632604599, 0.10427892208099365, -0.19334633648395538, -0.005319371819496155, 0.08282599598169327, 0.023504555225372314, 0.03946567326784134, 0.0035407328978180885, -0.03634254261851311, 0.055148303508758545, 0.02030518464744091, -0.08980578929185867, 0.14668866991996765, 0.0035520538222044706, -0.03514726087450981, -0.03927676007151604, -0.03267495706677437, 0.05703731253743172, 0.08045367896556854, -0.18214593827724457, -0.0733821839094162, -0.0838410034775734, -0.02458474040031433, 0.050523869693279266, 0.036679428070783615, 0.02738112211227417, 0.44813573360443115, 0.057562243193387985, 0.09003535658121109, -0.08811535686254501, 0.039806611835956573, 0.012785476632416248, -0.031281858682632446, 0.013625281862914562, 0.04725322127342224, 0.11279468983411789, 0.028284218162298203, 0.01669839769601822, 0.03680038824677467, 0.01938779093325138, 0.08824212104082108, -0.10939645022153854, -0.003965397831052542, 0.002614045049995184, 0.038018375635147095, 0.03672022372484207, 0.07190682739019394, 0.015936892479658127, -0.09583546966314316, -0.030848123133182526, -0.11166880279779434, 0.015594755299389362, -0.20979784429073334, -0.025905707851052284, -0.029619399458169937, 0.0003502996696624905, 0.09109684824943542, 0.04222718998789787, -0.04444896802306175, 0.035467714071273804, 0.03947039321064949, -0.0861397460103035, 0.0594942644238472, -0.014317752793431282, -0.07008631527423859, 0.13023322820663452, -0.1002996563911438, -0.3153233230113983, -0.08797995746135712, 0.05698639526963234, 0.05295826122164726, 0.06816939264535904, -0.05876303091645241, -0.09240786731243134, 0.03294730558991432, -0.06836386770009995, -0.0017794050509110093, 0.0037346978206187487, -0.051060982048511505, 0.07253886014223099, 0.08541567623615265, -0.014505518600344658, -0.08911184966564178, -0.006620637606829405, -0.041561197489500046, -0.124965138733387, 0.044060997664928436, -0.03760828450322151, 0.00007921225915197283, 0.18620672821998596, 0.03724536672234535, 0.06256633251905441, -0.06291008740663528, 0.07596296072006226, -0.09150096774101257, 0.0004740063741337508, 0.18428465723991394, -0.015377625823020935, -0.004100616089999676, -0.03996327146887779, -0.0259257685393095, -0.10829219967126846, 0.053985193371772766, -0.07330703735351562, -0.07349077612161636, -0.0023273853585124016, -0.07770214974880219, -0.0351552739739418, 0.0012160884216427803, 0.07817990332841873, 0.029699061065912247, -0.09635239094495773, 0.04920589178800583, 0.1298678070306778, 0.0931883230805397, 0.03626195341348648, 0.023981640115380287, 0.13739009201526642, -0.11230582743883133, 0.019063033163547516, -0.05148853361606598, -0.1041760966181755, -0.042787205427885056, -0.0714287981390953, 0.07368279993534088, 0.06034531816840172, -0.09970010071992874, 0.05144011229276657, 0.041872985661029816, 0.0883496031165123, 0.1373600959777832, -0.04213863983750343, -0.11244629323482513, -0.041393622756004333, -0.022004956379532814, -0.1777329444885254, 0.0341336652636528, 0.22155584394931793, 0.0073304991237819195, -0.10497386753559113, 0.07876885682344437, -0.005956185050308704, 0.11527370661497116, 0.031222699210047722, -0.278682678937912, 0.016931315883994102, 0.00203216471709311, 0.042359162122011185, -0.047676295042037964, 0.10937416553497314, 0.11747439950704575, -0.14421136677265167, -0.06650938838720322, -0.03273930773139, 0.044137366116046906, -0.15618287026882172, 0.036923591047525406, -0.12602220475673676, 0.06240779533982277, 0.050940994173288345, 0.05090156942605972, -0.2197665423154831, 0.06881614029407501, -0.0274215005338192, 0.06763827055692673, -0.062248338013887405, -0.01823522336781025, 0.04473711550235748, 0.025079863145947456, 0.14955177903175354, -0.014347962103784084, 0.14454017579555511, -0.09031219780445099, -0.11753576993942261, 0.0027052261866629124, 0.08532248437404633, 0.013173088431358337, 0.013580933213233948, 0.0026939227245748043, 0.041669201105833054, -0.02811569906771183, 0.17063532769680023, -0.08147624880075455, -0.022407781332731247, -0.06592555344104767, -0.018158966675400734, 0.2039334923028946, -0.12064731866121292, -0.10121093690395355, -0.11619500070810318, 0.08663272857666016, -0.04296411573886871, 0.08175522089004517, -0.020344657823443413, 0.049704354256391525, -0.02509051002562046, 0.007178863976150751, 0.09594997018575668, 0.01950966566801071, 0.08983828872442245, -0.09791163355112076, -0.019585272297263145, 0.13838915526866913, -0.037155888974666595, -0.036971647292375565, -0.019425252452492714, 0.11054370552301407, -0.0358734093606472, 0.08033111691474915, 0.03929615020751953, 0.03664831817150116, 0.03428546339273453, -0.039165496826171875, 0.10309428721666336, 0.10041618347167969, -0.06291446089744568, 0.03864621743559837, -0.07954532653093338, 0.26597461104393005, 0.040773067623376846, 0.07301845401525497, 0.28390514850616455, 0.19391325116157532, -0.03036464750766754, 0.10683353990316391, -0.017607249319553375, -0.024403288960456848, -0.2950931787490845, 0.0006976581644266844, 0.027765681967139244, 0.11812873929738998, 0.01744898222386837, -0.20587195456027985, -0.1211688369512558, -0.03560304269194603, -0.007791717536747456, 0.0310499370098114, -0.2441052496433258, -0.06442268192768097, 0.06107868626713753, 0.13779635727405548, 0.15878525376319885, -0.05917542055249214, -0.007856467738747597, 0.029358724132180214, 0.07593556493520737, 0.017292039468884468, -0.11598441749811172, 0.11550791561603546, 0.025637371465563774, -0.05708931386470795, 0.0267958827316761, -0.044003549963235855, 0.04214555397629738, -0.17736166715621948, 0.10933554917573929, -0.05924695357680321, -0.08421005308628082, 0.07140472531318665, -0.02217724733054638, -0.048552993685007095, 0.0789642184972763, 0.020652711391448975, -0.13173207640647888, 0.038154006004333496, 0.005618774797767401, 0.04346654564142227, -0.004941361024975777, -0.019811764359474182, -0.029163256287574768, 0.07706235349178314, -0.03806605935096741, 0.09605937451124191, 0.19590972363948822, -0.0573095865547657, 0.03974950686097145, 0.085201695561409, 0.09593135863542557, -0.05523005872964859, -0.0809539332985878, -0.03812742978334427, -0.005277194548398256, 0.0674438327550888, -0.08598461747169495, -0.019085103645920753, 0.07938229292631149, 0.015313901007175446, 0.14910826086997986, 0.14389736950397491, -0.08835655450820923, 0.11321785300970078, 0.10694554448127747, -0.11366690695285797, -0.08583837002515793, -0.02963297814130783, 0.0009990704711526632, 0.04910186678171158, -0.048617590218782425, 0.05932905897498131, -0.1035301461815834, 0.012819357216358185, 0.03532040864229202, 0.0038119733799248934, -0.09975302964448929, 0.009764863178133965, 0.08645275235176086, 0.06119582802057266, -0.0567571222782135, 0.09250631928443909, -0.0019178141374140978, -0.10868195444345474, 0.07241881638765335, 0.009918469935655594, -0.021528873592615128, -0.06352251768112183, 0.03211374953389168, 0.2370220273733139, 0.13945111632347107, -0.04336636886000633, -0.12396618723869324, -0.15508891642093658, 0.037849195301532745, 0.024356422945857048, 0.051251959055662155, 0.0062240250408649445, -0.06906022876501083, 0.01234503649175167, -0.04392383247613907, 0.005266309250146151, -0.05930564925074577, -0.047703344374895096, -0.12081446498632431, 0.1154373437166214, 0.053290288895368576, 0.11705748736858368, -0.0842847004532814, -0.07057584822177887, -0.1921386867761612, 0.09190598875284195, 0.041707299649715424, -0.05532265454530716, 0.06002674251794815, -0.030134430155158043, 0.017344338819384575, 0.11256659775972366, -0.051967836916446686, 0.008543911390006542, -0.09269233793020248, 0.03236149623990059, 0.03133073076605797, 0.04903566092252731, -0.004612727556377649, -0.017903391271829605, 0.04399999976158142, -0.05730267986655235, 0.07619527727365494, -0.07757602632045746, -0.033709146082401276, 0.0645759105682373, -0.16051416099071503, -0.054324716329574585, 0.08708633482456207, 0.013749903067946434, 0.02590017393231392, -0.05825240537524223, 0.019142305478453636, -0.05566488951444626, -0.04483235627412796, 0.01169554702937603, -0.05552767962217331, -0.011517677456140518, 0.05293213203549385, -0.05287189036607742, -0.040493328124284744, -0.06794002652168274, 0.061874233186244965, -0.07247710227966309, 0.09816460311412811, 0.031187955290079117, -0.10892423242330551, 0.07648903876543045, -0.037552736699581146, -0.0049397205002605915, -0.009439278393983841, 0.039307788014411926, 0.15598824620246887, -0.1606634259223938, 0.05345672369003296, -0.0484454482793808, 0.13272921741008759, 0.046888746321201324, -0.04458791762590408, -0.020207170397043228, 0.02469455823302269, -0.05549024045467377, 0.06932897865772247, 0.15877580642700195, 0.09880131483078003, 0.02571805939078331, 0.008134597912430763, 0.10187267512083054, 0.1060529574751854, 0.08136752992868423, 0.08394161611795425, -0.03428563475608826, -0.11287897825241089, 0.14338994026184082, 0.09748584777116776, 0.024613093584775925, 0.21077860891819, 0.17944025993347168, 0.03125298395752907, 0.03018142655491829, -0.06512103229761124, 0.17325744032859802, 0.061261482536792755, -0.08229418843984604, 0.014424329623579979, 0.03221147879958153, -0.049809664487838745, -0.047004032880067825, -0.09757380187511444, -0.029556652531027794, -0.24085633456707, 0.10851483792066574, -0.057250600308179855, -0.09750643372535706, 0.022772664204239845, 0.02990041859447956, -0.018839845433831215, 0.11280566453933716, -0.07735858112573624, 0.012980576604604721, 0.18577688932418823, -0.03825045004487038, -0.022322099655866623, -0.1633504331111908, -0.11154003441333771, -0.014046176336705685, -0.11750495433807373, 0.025494296103715897, 0.06305963546037674, 0.01117965579032898, 0.04399528726935387, 0.028923438861966133, -0.020834028720855713, 0.019218796864151955, -0.05903913825750351, -0.042673509567976, -0.01891910657286644, 0.02202831581234932, -0.09593231230974197, -0.03627033904194832, 0.12151803076267242, -0.03246605768799782, -0.08207374066114426, -0.006544890813529491, 0.07848484069108963, -0.042620159685611725, 0.09450104832649231, -0.07687012106180191, -0.03479038178920746, -0.06794454902410507, 0.268902063369751, 0.09388194978237152, -0.20183001458644867, 0.03341769427061081, -0.030470456928014755, 0.026735708117485046, -0.09215684235095978, 0.16250114142894745, 0.0899243950843811, 0.049168527126312256, -0.12686687707901, -0.003401300171390176, -0.09992645680904388, -0.0028723697178065777, -0.12552696466445923, -0.14725084602832794, 0.12093491852283478, -0.003848524997010827, -0.06547791510820389, 0.02844911813735962, -0.15909899771213531, 0.06585367769002914, 0.0978507474064827, -0.1514272391796112, -0.038227714598178864, -0.06086801365017891, 0.06072385236620903, 0.026465637609362602, 0.13005392253398895, -0.05080926790833473, 0.012067130766808987, -0.0656723901629448, -0.011309894733130932, -0.0000654291216051206, -0.017478201538324356, 0.001532604917883873, -0.09828947484493256, 0.05038110539317131, -0.0835796371102333, 0.12184429168701172, 0.05709611251950264, 0.005326167680323124, 0.008464806713163853, 0.0648408755660057, -0.02414623089134693, -0.10202058404684067, -0.01877439208328724, 0.033475372940301895, 0.03998998552560806, 0.010373802855610847, 0.034506846219301224, 0.0006507808575406671, 0.07714920490980148, -0.011413984932005405, -0.027285432443022728, -0.058209117501974106, 0.03936338797211647, -0.10441672056913376, 0.10461361706256866, 0.0013552121818065643, -0.02240127883851528, -0.010913821868598461, -0.05532446503639221, 0.045815300196409225, 0.04572062939405441, 0.029743505641818047, -0.05261747166514397, -0.09262793511152267, -0.021781492978334427, 0.023900283500552177, -0.11539579927921295, -0.18497975170612335, -0.0664035826921463, -0.15038692951202393, -0.01633414439857006, -0.0620744526386261, 0.08902198076248169, 0.13558129966259003, 0.030392181128263474, -0.04822919890284538, -0.12171997129917145, 0.025026977062225342, 0.13544774055480957, -0.03851630911231041, -0.07532322406768799 ]
null
null
transformers
## Exllama v2 Quantizations of Hercules-2.0-Mistral-7B Using <a href="https://github.com/turboderp/exllamav2/releases/tag/v0.0.13">turboderp's ExLlamaV2 v0.0.13</a> for quantization. <b>The "main" branch only contains the measurement.json, download one of the other branches for the model (see below)</b> Each branch contains an individual bits per weight, with the main one containing only the meaurement.json for further conversions. Original model: https://huggingface.co/Locutusque/Hercules-2.0-Mistral-7B | Branch | Bits | lm_head bits | VRAM (4k) | VRAM (16k) | VRAM (32k) | Description | | ----- | ---- | ------- | ------ | ------ | ------ | ------------ | | [8_0](https://huggingface.co/bartowski/Hercules-2.0-Mistral-7B-exl2/tree/8_0) | 8.0 | 8.0 | 8.4 GB | 9.8 GB | 11.8 GB | Maximum quality that ExLlamaV2 can produce, near unquantized performance. | | [6_5](https://huggingface.co/bartowski/Hercules-2.0-Mistral-7B-exl2/tree/6_5) | 6.5 | 8.0 | 7.2 GB | 8.6 GB | 10.6 GB | Very similar to 8.0, good tradeoff of size vs performance, **recommended**. | | [5_0](https://huggingface.co/bartowski/Hercules-2.0-Mistral-7B-exl2/tree/5_0) | 5.0 | 6.0 | 6.0 GB | 7.4 GB | 9.4 GB | Slightly lower quality vs 6.5, but usable on 8GB cards. | | [4_25](https://huggingface.co/bartowski/Hercules-2.0-Mistral-7B-exl2/tree/4_25) | 4.25 | 6.0 | 5.3 GB | 6.7 GB | 8.7 GB | GPTQ equivalent bits per weight, slightly higher quality. | | [3_5](https://huggingface.co/bartowski/Hercules-2.0-Mistral-7B-exl2/tree/3_5) | 3.5 | 6.0 | 4.7 GB | 6.1 GB | 8.1 GB | Lower quality, only use if you have to. | ## Download instructions With git: ```shell git clone --single-branch --branch 6_5 https://huggingface.co/bartowski/Hercules-2.0-Mistral-7B-exl2 Hercules-2.0-Mistral-7B-exl2-6_5 ``` With huggingface hub (credit to TheBloke for instructions): ```shell pip3 install huggingface-hub ``` To download the `main` (only useful if you only care about measurement.json) branch to a folder called `Hercules-2.0-Mistral-7B-exl2`: ```shell mkdir Hercules-2.0-Mistral-7B-exl2 huggingface-cli download bartowski/Hercules-2.0-Mistral-7B-exl2 --local-dir Hercules-2.0-Mistral-7B-exl2 --local-dir-use-symlinks False ``` To download from a different branch, add the `--revision` parameter: Linux: ```shell mkdir Hercules-2.0-Mistral-7B-exl2-6_5 huggingface-cli download bartowski/Hercules-2.0-Mistral-7B-exl2 --revision 6_5 --local-dir Hercules-2.0-Mistral-7B-exl2-6_5 --local-dir-use-symlinks False ``` Windows (which apparently doesn't like _ in folders sometimes?): ```shell mkdir Hercules-2.0-Mistral-7B-exl2-6.5 huggingface-cli download bartowski/Hercules-2.0-Mistral-7B-exl2 --revision 6_5 --local-dir Hercules-2.0-Mistral-7B-exl2-6.5 --local-dir-use-symlinks False ``` Want to support my work? Visit my ko-fi page here: https://ko-fi.com/bartowski
{"license": "apache-2.0", "library_name": "transformers", "tags": ["not-for-all-audiences", "chemistry", "math", "code", "physics"], "datasets": ["Locutusque/hercules-v2.0"], "widget": [{"example_title": "Derivative", "text": "<|im_start|>user\nIn this task, you are given a description of an object and its surroundings. Your task is to describe the motion of the object mentioned in the description using differential equations. The answer should be a mathematical expression in t. Description: A ball is thrown up at 30 ft/sec from 3 feet off the ground. Solve this step-by-step, explaining each step in detail.<|im_end|>\n<|im_start|>assistant\n", "output": {"text": "Let h(t) be the height of the ball at time t after it was thrown. We know that when the ball is thrown, it has an initial velocity v0 = 30 ft/sec and initial height y0 = 3 ft. We also know that the acceleration due to gravity g = -32 ft/sec^2.\n\nTo solve for the motion of the ball, we can use the equation for vertical motion with acceleration due to gravity:\n\nh(t) = v0 * t + (1/2) * g * t^2 + y0\n\nSubstituting the values given in the problem, we get:\n\ny(t) = (30)(t) + (1/2)(-32)(t)^2 + 3\n\nNow, let's simplify the equation:\n\ny(t) = 30t - 16t^2 + 3\n\nThis is the final solution for the height of the ball as a function of time."}}], "quantized_by": "bartowski", "pipeline_tag": "text-generation"}
text-generation
bartowski/Hercules-2.0-Mistral-7B-exl2
[ "transformers", "not-for-all-audiences", "chemistry", "math", "code", "physics", "text-generation", "dataset:Locutusque/hercules-v2.0", "license:apache-2.0", "endpoints_compatible", "region:us" ]
2024-02-10T14:00:39+00:00
[]
[]
TAGS #transformers #not-for-all-audiences #chemistry #math #code #physics #text-generation #dataset-Locutusque/hercules-v2.0 #license-apache-2.0 #endpoints_compatible #region-us
Exllama v2 Quantizations of Hercules-2.0-Mistral-7B --------------------------------------------------- Using <a href="URL ExLlamaV2 v0.0.13 for quantization. **The "main" branch only contains the URL, download one of the other branches for the model (see below)** Each branch contains an individual bits per weight, with the main one containing only the URL for further conversions. Original model: URL Download instructions --------------------- With git: With huggingface hub (credit to TheBloke for instructions): To download the 'main' (only useful if you only care about URL) branch to a folder called 'Hercules-2.0-Mistral-7B-exl2': To download from a different branch, add the '--revision' parameter: Linux: Windows (which apparently doesn't like \_ in folders sometimes?): Want to support my work? Visit my ko-fi page here: URL
[]
[ "TAGS\n#transformers #not-for-all-audiences #chemistry #math #code #physics #text-generation #dataset-Locutusque/hercules-v2.0 #license-apache-2.0 #endpoints_compatible #region-us \n" ]
[ 65 ]
[ "passage: TAGS\n#transformers #not-for-all-audiences #chemistry #math #code #physics #text-generation #dataset-Locutusque/hercules-v2.0 #license-apache-2.0 #endpoints_compatible #region-us \n" ]
[ -0.03898468241095543, 0.14885960519313812, -0.006139120552688837, 0.024796094745397568, 0.047405485063791275, 0.0263654887676239, 0.13011236488819122, 0.13548949360847473, 0.089565210044384, -0.0057469927705824375, 0.1375989317893982, 0.09521054476499557, -0.031148070469498634, 0.04698022082448006, -0.07349865138530731, -0.2105373740196228, 0.03510443866252899, 0.04703967645764351, -0.1287423074245453, 0.06371017545461655, 0.08241398632526398, -0.03986392170190811, 0.07080650329589844, -0.03682553768157959, -0.0007515348843298852, 0.012679186649620533, 0.0457131527364254, -0.12221192568540573, 0.10553260147571564, 0.04071391001343727, 0.09947817027568817, 0.11018701642751694, -0.09236235916614532, -0.2231162190437317, 0.04166834428906441, -0.009889267385005951, -0.064153291285038, 0.06391630321741104, -0.014083056710660458, -0.06435289233922958, 0.08889204263687134, -0.030131429433822632, -0.02353549934923649, 0.09978489577770233, -0.14450015127658844, -0.04482073709368706, -0.0886533260345459, -0.0914137214422226, -0.015534000471234322, 0.06342310458421707, 0.04055837169289589, 0.09584683179855347, -0.03235667571425438, 0.0212827417999506, 0.26256197690963745, -0.3527238965034485, -0.035137441009283066, 0.11073492467403412, 0.13963904976844788, 0.06866126507520676, -0.03952937573194504, 0.08794982731342316, 0.05703890323638916, 0.00937028881162405, 0.005651901476085186, -0.08275596797466278, -0.07156214118003845, 0.05920860171318054, -0.04237059876322746, 0.0035423238296061754, 0.2703133821487427, -0.014990394003689289, 0.015400199219584465, -0.018454227596521378, -0.11873193085193634, 0.0020089917816221714, -0.05164843797683716, 0.00037273234920576215, 0.059961576014757156, 0.08496447652578354, 0.01602286659181118, -0.04139549657702446, -0.109860859811306, 0.0013375661801546812, -0.11043939739465714, 0.08990471065044403, -0.07210589200258255, 0.04855377972126007, -0.1701536923646927, 0.04155177250504494, -0.05945280194282532, -0.1228460818529129, -0.060022734105587006, -0.12988074123859406, 0.12341742217540741, 0.00027315557235851884, -0.04755859822034836, -0.024449484422802925, 0.16452737152576447, 0.15570680797100067, 0.0454741045832634, 0.050146717578172684, 0.009139683097600937, 0.07665076851844788, -0.005791871342808008, -0.006289651617407799, 0.06281594932079315, -0.09476833045482635, 0.06942853331565857, -0.035304322838783264, 0.052948612719774246, -0.03976773843169212, -0.06394145637750626, -0.11465375870466232, -0.0502622053027153, 0.09211917221546173, 0.10443224012851715, 0.008463709615170956, -0.06608444452285767, 0.03545119985938072, 0.08265872299671173, -0.03343505412340164, -0.018724167719483376, 0.006289297249168158, 0.040076155215501785, 0.17733260989189148, 0.03734022378921509, 0.02011539787054062, 0.02168988063931465, -0.014097034931182861, -0.09269833564758301, -0.024843506515026093, -0.06754352897405624, 0.0241760965436697, 0.09894607216119766, -0.01688663475215435, 0.09108128398656845, -0.10773584246635437, -0.09446746855974197, 0.04678839445114136, 0.08706796169281006, -0.03748051077127457, -0.005691672675311565, 0.026165161281824112, -0.05238797143101692, 0.05774969980120659, -0.03613763302564621, -0.07542139291763306, -0.09038844704627991, 0.05520549789071083, -0.02764655463397503, 0.08592868596315384, -0.10161639004945755, 0.028063027188181877, -0.10734917968511581, 0.06004796549677849, -0.11692478507757187, -0.006882811896502972, -0.08661427348852158, 0.1163128986954689, -0.06228788197040558, -0.048487577587366104, -0.0027537725400179625, -0.017189666628837585, 0.011442018672823906, 0.2379685491323471, -0.16960401833057404, -0.042223960161209106, 0.15583829581737518, -0.08809873461723328, -0.27440884709358215, 0.038456033915281296, 0.008756960742175579, 0.03979284316301346, 0.04695212468504906, 0.11031851917505264, 0.053707752376794815, -0.07025060802698135, -0.041531432420015335, 0.1323317438364029, 0.011412288062274456, -0.08563104271888733, 0.055121418088674545, 0.02814723737537861, -0.04946567863225937, 0.057779233902692795, 0.016374941915273666, 0.03225664049386978, -0.024217305704951286, -0.08859293907880783, -0.07811122387647629, -0.022021543234586716, 0.06368811428546906, -0.0018084748880937696, 0.0736207365989685, -0.0644674301147461, -0.01859462261199951, 0.019394079223275185, 0.03450379893183708, -0.0030639960896223783, 0.06679825484752655, -0.03611605614423752, 0.10594717413187027, 0.020430253818631172, 0.03280143439769745, -0.12847891449928284, -0.031235221773386, -0.029879529029130936, 0.009129582904279232, -0.015845943242311478, 0.05213043466210365, 0.10229214280843735, -0.06635835021734238, -0.06187159940600395, 0.0172573272138834, 0.006982370745390654, 0.047838855534791946, -0.011646597646176815, -0.1423289179801941, 0.04099544137716293, -0.05849260464310646, 0.02156226709485054, -0.07263565808534622, 0.03783416375517845, 0.08028559386730194, 0.06236438453197479, 0.016302285715937614, 0.042613208293914795, -0.0215340256690979, -0.024909181520342827, -0.0648607462644577, -0.014729449525475502, 0.059328120201826096, 0.006624090019613504, -0.11462210863828659, 0.015612679533660412, -0.03876514360308647, 0.24046748876571655, 0.20572111010551453, -0.21796686947345734, 0.07639659196138382, -0.04209073260426521, -0.008302230387926102, 0.042180828750133514, 0.07113799452781677, 0.06057911366224289, -0.03813200071454048, 0.007439527660608292, 0.05721645429730415, -0.03292534872889519, 0.03955749049782753, 0.017930055037140846, -0.0761740580201149, -0.04141807556152344, 0.03197783604264259, 0.10446391999721527, -0.017734482884407043, 0.1471070647239685, 0.3254410922527313, -0.009760911576449871, 0.08975622057914734, -0.03413073346018791, -0.033183276653289795, 0.021708300337195396, -0.0604463592171669, -0.0016061464557424188, 0.03202101215720177, -0.1884377896785736, 0.013569623231887817, 0.06836511194705963, -0.014115802943706512, 0.048520613461732864, -0.09157309681177139, -0.08167454600334167, -0.003310841042548418, -0.011630700901150703, -0.12361020594835281, 0.038813069462776184, -0.050466496497392654, 0.10578488558530807, -0.0020163431763648987, -0.002921830164268613, 0.09337884187698364, -0.0014636320993304253, -0.08975035697221756, 0.1849299669265747, -0.10133028030395508, -0.08015768229961395, -0.12186983227729797, -0.03034648299217224, -0.017686942592263222, -0.0004207154852338135, 0.0888325646519661, -0.053723327815532684, -0.05848651006817818, -0.017908789217472076, 0.022380100563168526, -0.030197082087397575, 0.03327500820159912, 0.0019492593128234148, 0.07360592484474182, -0.025803452357649803, -0.10188641399145126, -0.033288124948740005, 0.0218382328748703, -0.024035966023802757, 0.10152359306812286, -0.11273670196533203, 0.09836490452289581, 0.09185130894184113, 0.029233912006020546, -0.00348053895868361, -0.031567398458719254, 0.1883401721715927, -0.09136959165334702, 0.013060159049928188, 0.20819170773029327, 0.05604531615972519, 0.05058712884783745, 0.10887761414051056, 0.06884060055017471, -0.07282412797212601, 0.014140898361802101, -0.047174058854579926, -0.09916834533214569, -0.2951529622077942, -0.06920307129621506, -0.11842244118452072, 0.07790680974721909, -0.04648034647107124, 0.035916928201913834, 0.09450144320726395, 0.09590055048465729, -0.0002784634125418961, -0.011067814193665981, 0.002997200470417738, 0.05560077354311943, 0.22479422390460968, 0.03611820191144943, 0.07456955313682556, -0.10958190262317657, -0.007329231593757868, 0.11150921881198883, 0.11015201359987259, 0.18344898521900177, 0.11065850406885147, 0.1649140864610672, 0.051255591213703156, 0.06190057471394539, 0.07720399647951126, 0.1169469952583313, 0.06471294164657593, -0.012733051553368568, -0.04058726504445076, -0.04469196870923042, -0.04307899624109268, 0.08269603550434113, -0.09316959977149963, -0.11854472011327744, -0.03365934640169144, -0.06227672100067139, 0.10548032075166702, 0.17301113903522491, 0.007939872331917286, -0.13168925046920776, -0.03198797628283501, 0.16252216696739197, 0.041286565363407135, -0.06361941993236542, 0.10170423239469528, -0.0661158636212349, -0.05298426374793053, 0.1388292759656906, -0.07752726972103119, 0.08720064163208008, 0.08795051276683807, 0.044576097279787064, -0.005034870933741331, -0.17596109211444855, 0.03476062789559364, 0.08212129026651382, -0.3255690336227417, 0.1850433498620987, 0.006845721043646336, 0.01599547453224659, -0.04083386808633804, 0.013448956422507763, 0.02289990894496441, 0.22654534876346588, 0.18150071799755096, 0.0050564119592309, -0.028104130178689957, -0.0037091358099132776, -0.01973477564752102, 0.06683117896318436, 0.03304635360836983, 0.016491642221808434, 0.00022161395463626832, -0.030362512916326523, -0.0006580979097634554, 0.03440156206488609, 0.04405944421887398, -0.06521853804588318, -0.15878459811210632, 0.10440876334905624, -0.03878719359636307, 0.04122660309076309, -0.034316759556531906, -0.053742554038763046, -0.0013691229978576303, 0.127048060297966, -0.08056597411632538, -0.058668170124292374, -0.0713694840669632, -0.15691688656806946, 0.11315252631902695, -0.05199673771858215, 0.06299317628145218, -0.04578890651464462, -0.013892180286347866, -0.0638163760304451, -0.20944707095623016, 0.1656537503004074, -0.09991069883108139, -0.07966241985559464, -0.02228919044137001, 0.08486542850732803, -0.07150374352931976, 0.06978277862071991, -0.007083714008331299, 0.06621692329645157, -0.1222265213727951, -0.12188329547643661, -0.01121850311756134, -0.022451195865869522, 0.01182085182517767, -0.09445326775312424, -0.004371081478893757, -0.035336438566446304, 0.057408954948186874, -0.007093760184943676, 0.1523449569940567, 0.28523463010787964, -0.07257364690303802, 0.1225123330950737, 0.1796312928199768, -0.08958837389945984, -0.24901635944843292, -0.03831484913825989, -0.20458528399467468, -0.06518949568271637, -0.02689766325056553, -0.1995759755373001, 0.15322907269001007, 0.0016423857305198908, -0.049159932881593704, 0.13852863013744354, -0.1988477259874344, -0.0922384187579155, 0.07977709919214249, -0.01759498007595539, 0.3292481601238251, -0.21457086503505707, -0.05667589604854584, -0.08040543645620346, -0.2661356031894684, 0.17101167142391205, -0.18406033515930176, 0.05564919859170914, -0.037622448056936264, 0.019569741562008858, 0.0003766227455344051, -0.08994854241609573, 0.17861095070838928, -0.01863596960902214, 0.02284722402691841, -0.0406169556081295, -0.09619870781898499, 0.11308111995458603, -0.019031427800655365, 0.003801709972321987, -0.07814636826515198, 0.02468198724091053, -0.055510133504867554, -0.02825363539159298, -0.06376361846923828, 0.11582595109939575, -0.021569661796092987, -0.07872643321752548, -0.10270506143569946, -0.02189725823700428, 0.025005651637911797, 0.0009406518074683845, 0.3093308210372925, -0.019981222227215767, 0.11417784541845322, 0.11615028977394104, 0.20803950726985931, -0.21219293773174286, 0.026984605938196182, 0.06126263365149498, -0.03784889727830887, 0.04689132049679756, -0.24720360338687897, 0.039815764874219894, 0.13484811782836914, -0.02997649647295475, -0.00931258499622345, 0.10322320461273193, 0.002760291565209627, 0.04125746339559555, 0.09015265107154846, -0.19508002698421478, -0.09450241923332214, -0.030451426282525063, 0.050323229283094406, -0.040685102343559265, 0.09029831737279892, 0.10216040909290314, -0.014856773428618908, -0.021737953647971153, -0.0265969205647707, 0.03804972022771835, -0.0670345351099968, 0.011370010673999786, 0.043014202266931534, -0.0011285964865237474, -0.07673116028308868, 0.016588080674409866, -0.003659680252894759, -0.23734335601329803, -0.04839026555418968, 0.0873866006731987, -0.09362699836492538, -0.11216980963945389, -0.005026089958846569, 0.038747526705265045, -0.16970200836658478, -0.014571504667401314, -0.11787547171115875, -0.14557263255119324, 0.03704971447587013, 0.17273740470409393, 0.08698093891143799, 0.05401809141039848, -0.031145406886935234, -0.005298679228872061, 0.07740388065576553, -0.02837272919714451, -0.06797929108142853, 0.036406248807907104, -0.054113034158945084, -0.09893645346164703, -0.03216923773288727, 0.06975740194320679, -0.08126850426197052, -0.007657230831682682, -0.0847361832857132, 0.01280779205262661, -0.14634205400943756, 0.0027000908739864826, -0.1612677127122879, -0.02393748052418232, -0.052939705550670624, -0.05103606730699539, -0.08263653516769409, -0.022685088217258453, -0.0818927213549614, 0.017558420076966286, -0.05366358160972595, 0.08109603077173233, -0.131059929728508, -0.0025656498037278652, 0.09064270555973053, -0.07909911870956421, 0.07146193087100983, 0.07946456968784332, -0.051706716418266296, 0.05939915031194687, -0.1712227761745453, -0.05155906453728676, 0.06990131735801697, 0.06534431874752045, -0.010276399552822113, -0.01354359369724989, -0.009720786474645138, 0.08068938553333282, 0.04700155928730965, 0.01822596974670887, -0.04633794352412224, -0.094783715903759, -0.051395684480667114, -0.012603911571204662, -0.12171559035778046, -0.02924659289419651, -0.045178528875112534, 0.23872917890548706, -0.014580871909856796, 0.14017270505428314, 0.0024330378510057926, 0.02061963826417923, -0.1209823489189148, 0.03335350379347801, 0.0016922648064792156, -0.12846045196056366, -0.07026173174381256, -0.04330871254205704, -0.0006040805601514876, -0.02540675923228264, 0.23091596364974976, -0.04207148775458336, 0.024351857602596283, 0.08002245426177979, 0.056336771696805954, 0.035757701843976974, 0.06127987802028656, 0.2842041254043579, 0.11002053320407867, 0.00918260496109724, -0.07900860905647278, 0.05176178365945816, 0.023745877668261528, 0.02722996287047863, 0.14109236001968384, 0.11443346738815308, 0.002432028530165553, 0.10954196006059647, -0.04271410405635834, 0.03886032477021217, -0.05387790873646736, -0.052077729254961014, 0.016823604702949524, 0.027417028322815895, -0.023485006764531136, 0.11120080202817917, 0.14344798028469086, -0.06377323716878891, 0.02984093874692917, -0.13053537905216217, -0.05228274315595627, -0.14396600425243378, 0.014157257042825222, -0.08572547882795334, -0.15696337819099426, -0.019125506281852722, -0.11402327567338943, 0.019940881058573723, 0.10137826949357986, 0.0031506677623838186, -0.04828263819217682, 0.07062707841396332, 0.059830665588378906, -0.01646622084081173, 0.04843776300549507, 0.0029532865155488253, -0.02417079173028469, -0.02750597707927227, -0.032031264156103134, -0.08157991617918015, -0.038751423358917236, -0.06538381427526474, 0.07296931743621826, -0.0454360768198967, 0.03342238813638687, -0.09969276189804077, -0.08625348657369614, -0.03906610608100891, 0.05472178757190704, -0.005943180527538061, 0.11662722378969193, -0.013879378326237202, 0.006115593947470188, 0.09845831245183945, 0.21571248769760132, -0.07966732233762741, -0.06518608331680298, -0.06008845567703247, 0.041654158383607864, 0.0067101819440722466, 0.10270795226097107, -0.01203032024204731, -0.031020184978842735, -0.0490044504404068, 0.1802278310060501, 0.25360924005508423, -0.04759620130062103, 0.007601925637573004, 0.003920952323824167, 0.04358215630054474, 0.03949980437755585, 0.0987459048628807, 0.09934666007757187, 0.16390351951122284, -0.13244155049324036, -0.03336348012089729, -0.07149462401866913, -0.0021888199262320995, -0.11360620707273483, 0.13749021291732788, 0.019428974017500877, -0.04521840065717697, -0.008203605189919472, 0.11087816208600998, -0.11983142793178558, 0.1113620400428772, 0.0008097956306301057, -0.1955053061246872, -0.021390685811638832, -0.03929872810840607, 0.11408708989620209, -0.03376585617661476, 0.0609716959297657, -0.027247538790106773, -0.08124659210443497, 0.11584337055683136, 0.052855562418699265, -0.20201510190963745, 0.014214148744940758, 0.15002062916755676, -0.08566979318857193, 0.11348503082990646, 0.00037455279380083084, 0.008116417564451694, 0.0955355167388916, 0.05101977661252022, -0.06394067406654358, 0.028239000588655472, 0.10461446642875671, -0.03983293101191521, -0.02630706876516342, -0.06539075821638107, -0.031140724197030067, 0.004610791802406311, 0.07887431234121323, -0.03626743331551552, 0.03421781212091446, 0.11300761997699738, -0.08215109258890152, -0.038892291486263275, 0.004332699812948704, -0.08471249788999557, 0.04362901672720909, -0.02776978351175785, -0.05391142517328262, -0.03752508386969566, -0.02347113937139511, 0.021308794617652893, 0.020216144621372223, -0.07304832339286804, -0.09958919882774353, -0.061499450355768204, -0.06350287050008774, 0.09736095368862152, 0.07686607539653778, -0.13097523152828217, -0.040785107761621475, -0.054415859282016754, 0.09300439059734344, -0.1268543004989624, 0.021896304562687874, 0.058148343116045, -0.06646484136581421, -0.019326871261000633, -0.0915583074092865, 0.022571370005607605, 0.08994545787572861, -0.09474721550941467, -0.10311421006917953 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # Donut_trainthai This model is a fine-tuned version of [naver-clova-ix/donut-base](https://huggingface.co/naver-clova-ix/donut-base) on the imagefolder dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 2 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results ### Framework versions - Transformers 4.38.0.dev0 - Pytorch 2.1.2+cpu - Datasets 2.16.1 - Tokenizers 0.15.1
{"license": "mit", "tags": ["generated_from_trainer"], "datasets": ["imagefolder"], "base_model": "naver-clova-ix/donut-base", "model-index": [{"name": "Donut_trainthai", "results": []}]}
null
suthawadee/Donut_trainthai
[ "transformers", "tensorboard", "safetensors", "vision-encoder-decoder", "generated_from_trainer", "dataset:imagefolder", "base_model:naver-clova-ix/donut-base", "license:mit", "endpoints_compatible", "region:us" ]
2024-02-10T14:03:54+00:00
[]
[]
TAGS #transformers #tensorboard #safetensors #vision-encoder-decoder #generated_from_trainer #dataset-imagefolder #base_model-naver-clova-ix/donut-base #license-mit #endpoints_compatible #region-us
# Donut_trainthai This model is a fine-tuned version of naver-clova-ix/donut-base on the imagefolder dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 2 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results ### Framework versions - Transformers 4.38.0.dev0 - Pytorch 2.1.2+cpu - Datasets 2.16.1 - Tokenizers 0.15.1
[ "# Donut_trainthai\n\nThis model is a fine-tuned version of naver-clova-ix/donut-base on the imagefolder dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 2\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3", "### Training results", "### Framework versions\n\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.2+cpu\n- Datasets 2.16.1\n- Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #tensorboard #safetensors #vision-encoder-decoder #generated_from_trainer #dataset-imagefolder #base_model-naver-clova-ix/donut-base #license-mit #endpoints_compatible #region-us \n", "# Donut_trainthai\n\nThis model is a fine-tuned version of naver-clova-ix/donut-base on the imagefolder dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 2\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3", "### Training results", "### Framework versions\n\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.2+cpu\n- Datasets 2.16.1\n- Tokenizers 0.15.1" ]
[ 70, 37, 6, 12, 8, 3, 90, 4, 38 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #vision-encoder-decoder #generated_from_trainer #dataset-imagefolder #base_model-naver-clova-ix/donut-base #license-mit #endpoints_compatible #region-us \n# Donut_trainthai\n\nThis model is a fine-tuned version of naver-clova-ix/donut-base on the imagefolder dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 2\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3### Training results### Framework versions\n\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.2+cpu\n- Datasets 2.16.1\n- Tokenizers 0.15.1" ]
[ -0.0816536471247673, 0.12812598049640656, -0.0009826552122831345, 0.09274964779615402, 0.14790134131908417, 0.005476665683090687, 0.06247246637940407, 0.08807101100683212, -0.12135262042284012, 0.06918962299823761, 0.0689961165189743, 0.026871267706155777, 0.03752180561423302, 0.15846388041973114, -0.028158649802207947, -0.2608456015586853, 0.007719497662037611, 0.016873126849532127, -0.064204141497612, 0.09175697714090347, 0.05983170494437218, -0.10132712125778198, 0.07440747320652008, -0.024045191705226898, -0.16021470725536346, 0.02875298447906971, -0.05846061185002327, -0.053697358816862106, 0.06127455458045006, 0.003786807879805565, 0.11641668528318405, 0.021329304203391075, 0.1503070592880249, -0.19694659113883972, 0.00012179543409729376, 0.09826074540615082, 0.039164554327726364, 0.06624791771173477, 0.052707184106111526, 0.0246878694742918, 0.10459443181753159, -0.19955609738826752, 0.111156165599823, 0.03052148036658764, -0.056988053023815155, -0.12520425021648407, -0.06621968001127243, 0.11482115834951401, 0.11165311932563782, 0.11826426535844803, 0.00820094533264637, 0.12368965893983841, -0.03302513808012009, 0.09006504714488983, 0.15903139114379883, -0.18917050957679749, -0.07317324727773666, 0.12791915237903595, 0.030717190355062485, 0.07788525521755219, -0.10708602517843246, 0.019400974735617638, 0.07793499529361725, 0.020711423829197884, 0.08373014628887177, -0.021666020154953003, -0.05978788435459137, -0.028445713222026825, -0.15009118616580963, -0.022948451340198517, 0.1373649537563324, 0.06813952326774597, -0.04855717346072197, -0.07185295224189758, -0.0489015057682991, -0.11620690673589706, -0.023579705506563187, -0.06438356637954712, 0.0508299395442009, -0.05006873607635498, -0.07235278934240341, -0.098824143409729, -0.08275638520717621, -0.06391813606023788, 0.00650416873395443, 0.061254747211933136, 0.04361309856176376, -0.002151456428691745, -0.003427689429372549, 0.1047704666852951, -0.017495591193437576, -0.10430874675512314, -0.03052603267133236, 0.0006491434760391712, -0.085554338991642, -0.08193322271108627, 0.00017661834135651588, -0.029671406373381615, -0.02807151898741722, 0.09375406056642532, -0.08950041234493256, 0.08059760928153992, -0.04662876948714256, 0.006468453444540501, -0.04160843789577484, 0.13783767819404602, -0.04069843888282776, 0.003300265409052372, -0.012117895297706127, 0.0773945227265358, 0.005165190901607275, -0.009524391032755375, -0.09135891497135162, -0.018102264031767845, 0.08035899698734283, 0.05867375433444977, -0.04742563143372536, 0.053480569273233414, 0.004193809349089861, -0.021957796066999435, 0.02041321061551571, -0.11799508333206177, 0.06583372503519058, -0.01502388995140791, -0.09074604511260986, 0.00870926957577467, 0.06620447337627411, -0.003102555638179183, -0.03762555494904518, 0.09399324655532837, -0.05936434119939804, 0.007660874631255865, -0.0982884094119072, -0.05750500038266182, 0.031089523807168007, -0.07787440717220306, -0.01785150356590748, -0.0873088464140892, -0.22964578866958618, -0.049190834164619446, 0.04045509919524193, -0.04547114297747612, -0.02881345897912979, -0.06110364571213722, -0.09405636787414551, -0.010519652627408504, 0.017034152522683144, 0.10825250297784805, -0.03561263903975487, 0.058467503637075424, -0.010608410462737083, 0.01662854477763176, 0.039410024881362915, 0.031885307282209396, -0.08244242519140244, 0.034560807049274445, -0.15994013845920563, 0.07679684460163116, -0.08739577978849411, 0.04505513980984688, -0.12056997418403625, -0.11738572269678116, -0.051039230078458786, -0.021430466324090958, 0.018955299630761147, 0.1206519603729248, -0.17971765995025635, -0.00765744224190712, 0.11654551327228546, -0.09035700559616089, -0.0555102601647377, 0.10383883118629456, -0.039174698293209076, 0.008402310311794281, 0.07137079536914825, 0.13066880404949188, 0.11419250071048737, -0.1629614531993866, -0.0020307092927396297, -0.00949179194867611, 0.028799161314964294, 0.008566985838115215, 0.04620324820280075, -0.003851366462185979, 0.025788545608520508, -0.009760109707713127, -0.09165174514055252, -0.0037134375888854265, -0.07979229837656021, -0.07898381352424622, -0.05260150507092476, -0.08264359086751938, 0.02466772124171257, 0.040623437613248825, 0.04125192016363144, -0.06274345517158508, -0.10942036658525467, 0.10451558232307434, 0.109886534512043, -0.07063588500022888, 0.007157417014241219, -0.052292514592409134, -0.013631106354296207, -0.056837618350982666, -0.025275355204939842, -0.15777170658111572, -0.12514570355415344, 0.017005393281579018, -0.10427342355251312, 0.0612286776304245, 0.009288208559155464, 0.05807331204414368, 0.09503864496946335, -0.04091276228427887, -0.03366808220744133, -0.08602892607450485, 0.0214758962392807, -0.07906446605920792, -0.17476631700992584, -0.04183505102992058, -0.014723340980708599, 0.1311081498861313, -0.2801164984703064, 0.03506477177143097, 0.006993383169174194, 0.14062821865081787, 0.032608870416879654, -0.05578579753637314, 0.007907640188932419, 0.05827530845999718, 0.006924065761268139, -0.12122172862291336, 0.031010333448648453, -0.005621870514005423, -0.05849633738398552, -0.07814107090234756, -0.15910224616527557, 0.012201972305774689, 0.08903368562459946, 0.0428159274160862, -0.1069234237074852, 0.003862061770632863, -0.05838115140795708, -0.0485120452940464, -0.09394404292106628, -0.0030285923276096582, 0.16618438065052032, -0.0010005601216107607, 0.12957994639873505, -0.0311999823898077, -0.05123784765601158, 0.017060888931155205, 0.015389262698590755, -0.06352901458740234, 0.059681009501218796, 0.09130339324474335, -0.1149873360991478, 0.10563630610704422, 0.07320147007703781, -0.046204518526792526, 0.1589435636997223, -0.03727743774652481, -0.07480713725090027, -0.01683332398533821, 0.025138726457953453, -0.004760392010211945, 0.11356057226657867, -0.07261970639228821, 0.005610120948404074, -0.0029362495988607407, 0.009163382463157177, 0.039179522544145584, -0.17629709839820862, 0.0065828608348965645, 0.059642937034368515, -0.042011458426713943, 0.04352039098739624, -0.046040549874305725, 0.012731596827507019, 0.06390590220689774, 0.011621087789535522, -0.027740277349948883, 0.02050151117146015, 0.006519525311887264, -0.08306928724050522, 0.14988566935062408, -0.09505823254585266, -0.14387744665145874, -0.12634584307670593, 0.08090415596961975, -0.04346442222595215, -0.014308962970972061, 0.013412829488515854, -0.07619822025299072, -0.06076118350028992, -0.11636282503604889, -0.0175393745303154, -0.06233655661344528, -0.019001150503754616, 0.07749376446008682, 0.02518402971327305, 0.05768394097685814, -0.11085128784179688, 0.008827383629977703, -0.002857774496078491, -0.12525971233844757, -0.022502318024635315, 0.0361148901283741, 0.12738250195980072, 0.13362611830234528, -0.03392728790640831, 0.02646179497241974, -0.03857830539345741, 0.14465413987636566, -0.08187408000230789, 0.023731254041194916, 0.11055110394954681, 0.023249737918376923, 0.045100290328264236, 0.11691812425851822, 0.017736928537487984, -0.08316262066364288, 0.03924939036369324, 0.08153826743364334, -0.04258403182029724, -0.18785686790943146, -0.04951527714729309, -0.026594480499625206, -0.02896076627075672, 0.10921835899353027, 0.0549045167863369, 0.07404450327157974, 0.04866861179471016, -0.01686885766685009, 0.08111633360385895, 0.0047670528292655945, 0.09209250658750534, 0.10082370787858963, 0.030410997569561005, 0.0607636421918869, -0.04390048235654831, -0.047423142939805984, 0.04429518058896065, -0.0038811855483800173, 0.2621518075466156, -0.018209142610430717, 0.07631950080394745, 0.01837949827313423, 0.09346304833889008, -0.01816806010901928, 0.014064505696296692, 0.03864314407110214, -0.0004884809022769332, 0.02608415298163891, -0.07579656690359116, 0.011357801035046577, 0.03236328437924385, -0.032001107931137085, 0.04068192467093468, -0.08176230639219284, 0.06810984760522842, 0.021291935816407204, 0.2220853865146637, 0.0073527623899281025, -0.32856062054634094, -0.07043591886758804, 0.006104986183345318, 0.014226812869310379, -0.048012424260377884, 0.0033355653285980225, 0.14897221326828003, -0.12000349909067154, 0.07466765493154526, -0.05797332152724266, 0.08562059700489044, -0.05290994048118591, 0.009021461009979248, 0.06486091017723083, 0.11289732158184052, -0.004182647913694382, 0.07593363523483276, -0.19618922472000122, 0.1985234171152115, 0.011285045184195042, 0.14340217411518097, -0.0495610237121582, 0.030399765819311142, 0.03187576308846474, 0.12666065990924835, 0.1325961798429489, -0.017447572201490402, -0.0823763981461525, -0.20362922549247742, -0.07562196999788284, 0.01844392903149128, 0.11356663703918457, -0.020904280245304108, 0.06385978311300278, -0.03284558653831482, -0.02210555598139763, 0.056981172412633896, -0.04473740607500076, -0.22266645729541779, -0.10381455719470978, -0.007499754894524813, 0.012311818078160286, -0.03905501589179039, -0.07122927159070969, -0.10371482372283936, -0.044458579272031784, 0.15168017148971558, -0.022431012243032455, -0.021070262417197227, -0.1432788074016571, 0.11647364497184753, 0.08644948154687881, -0.05544855445623398, 0.018383312970399857, 0.03937050700187683, 0.15053720772266388, 0.036848053336143494, -0.09355521947145462, 0.04537831246852875, -0.07436773926019669, -0.14944997429847717, -0.03840174525976181, 0.09808894991874695, 0.051238760352134705, 0.037567056715488434, -0.0008614832768216729, 0.03066732920706272, 0.06702494621276855, -0.08829847723245621, 0.011013462208211422, 0.07733801007270813, 0.07503893971443176, 0.07829143106937408, -0.06938773393630981, -0.0349300242960453, -0.028684910386800766, 0.0033428736496716738, 0.09328502416610718, 0.15608663856983185, -0.0728173553943634, 0.02711460553109646, 0.050054263323545456, -0.09101346135139465, -0.18354089558124542, 0.12961234152317047, 0.07392972707748413, -0.0012627020478248596, 0.023883337154984474, -0.162794291973114, 0.11342469602823257, 0.1393996924161911, -0.021747568622231483, 0.051787808537483215, -0.33697402477264404, -0.13349281251430511, 0.07401954382658005, 0.14038962125778198, -0.05561281740665436, -0.14092394709587097, -0.03382313251495361, -0.018062643706798553, -0.15052783489227295, 0.1051076278090477, -0.1031704917550087, 0.09190206229686737, 0.016717756167054176, 0.042256806045770645, 0.024907078593969345, -0.041036199778318405, 0.12293676286935806, 0.03154231607913971, 0.09680978953838348, -0.07887406647205353, 0.03481999784708023, 0.09547153115272522, -0.08081004023551941, 0.08981682360172272, -0.0009597906609997153, 0.055481307208538055, -0.08035506308078766, -0.019397752359509468, -0.04394792392849922, 0.0757591724395752, -0.055694811046123505, -0.06019022315740585, -0.05019238591194153, 0.06584806740283966, 0.06395898759365082, -0.037357110530138016, 0.06425803899765015, 0.03380207344889641, 0.08587461709976196, 0.0949234738945961, 0.058385834097862244, 0.01616501435637474, -0.11084972321987152, -0.005064621567726135, -0.020076923072338104, 0.06822377443313599, -0.12049712240695953, 0.015164291486144066, 0.1341867595911026, 0.045642659068107605, 0.14915387332439423, 0.03708085045218468, -0.0461193323135376, -0.009818710386753082, 0.01768522337079048, -0.07711470872163773, -0.1738705188035965, -0.01379301492124796, -0.08400154858827591, -0.07920362055301666, 0.017790615558624268, 0.09737817943096161, -0.11102347075939178, 0.030986489728093147, -0.02936999686062336, 0.022252388298511505, -0.01632230542600155, 0.17725077271461487, 0.04624531418085098, 0.05786201357841492, -0.05733349546790123, 0.12488792091608047, 0.08542267978191376, -0.11301679909229279, 0.05961117148399353, 0.09069715440273285, -0.07349088042974472, -0.032987598329782486, 0.12266550213098526, 0.19262158870697021, -0.02510376088321209, -0.044091422110795975, -0.06754685938358307, -0.08801548182964325, 0.044954001903533936, 0.11911714822053909, 0.03717627003788948, -0.01650186814367771, -0.03355081379413605, 0.04446687176823616, -0.1808656007051468, 0.09707291424274445, 0.027130529284477234, 0.08744480460882187, -0.1932506114244461, 0.13014353811740875, 0.029193218797445297, 0.03313128277659416, -0.023441005498170853, 0.01510642096400261, -0.07750698179006577, -0.011626409366726875, -0.06783963739871979, 0.006904487498104572, -0.02819884568452835, -0.0009070029482245445, -0.008594646118581295, -0.03242192789912224, -0.04109552502632141, 0.06976945698261261, -0.04673261195421219, -0.07455084472894669, 0.015331814996898174, 0.058221861720085144, -0.1441490650177002, -0.007878188975155354, 0.010039004497230053, -0.08293388038873672, 0.01750847138464451, 0.016169829294085503, 0.01445255521684885, 0.019092176109552383, -0.1666945219039917, 0.007807841524481773, 0.06695810705423355, 0.02065216936171055, 0.051795221865177155, -0.06124356761574745, -0.009506729431450367, -0.017693908885121346, 0.05024262145161629, 0.017333611845970154, 0.07450703531503677, -0.116572305560112, -0.03919966146349907, -0.07504929602146149, -0.03495200723409653, -0.04983489215373993, 0.06452571600675583, 0.09859781712293625, 0.027693413197994232, 0.14862623810768127, -0.09286560863256454, 0.04074862599372864, -0.1783110797405243, -0.04972832649946213, 0.007442713715136051, -0.028284652158617973, 0.0013934681192040443, -0.01952281966805458, 0.07568421214818954, -0.0574682392179966, 0.06398099660873413, 0.022269511595368385, 0.11503145098686218, 0.052950166165828705, -0.04672307148575783, -0.019769279286265373, 0.03185616806149483, 0.16536107659339905, 0.03950610011816025, -0.02939569763839245, 0.1042579934000969, 0.0005698559689335525, 0.09937916696071625, 0.03509242832660675, 0.18768402934074402, 0.1544203758239746, -0.09138426929712296, 0.06298302114009857, 0.09712023288011551, -0.0826134979724884, -0.16990122199058533, 0.08934291452169418, -0.043812889605760574, 0.1300644427537918, -0.06374382227659225, 0.0843268632888794, 0.11187276989221573, -0.1820427030324936, 0.03136634826660156, -0.03351826220750809, -0.11204259842634201, -0.08353900909423828, -0.10289133340120316, -0.08398314565420151, -0.10653582215309143, 0.03587856516242027, -0.09826916456222534, 0.013009593822062016, 0.08831588923931122, 0.018588988110423088, 0.021986762061715126, 0.214424267411232, -0.05763157457113266, 0.027231762185692787, 0.0679093450307846, 0.008105848915874958, -0.03171994164586067, -0.07657891511917114, -0.04663645476102829, 0.06087101250886917, 0.007249628193676472, 0.0726127102971077, -0.049254387617111206, -0.0016685682348906994, 0.05130210146307945, -0.010269350372254848, -0.06363482028245926, 0.033434901386499405, 0.02118467539548874, 0.032485418021678925, 0.016262473538517952, 0.046679068356752396, -0.03211116045713425, -0.03929479792714119, 0.27131134271621704, -0.06828289479017258, -0.09891381114721298, -0.11080686002969742, 0.13407213985919952, 0.06697005778551102, -0.019032372161746025, 0.06650646775960922, -0.13340716063976288, 0.014008378610014915, 0.13393434882164001, 0.14618195593357086, -0.014791113324463367, -0.009229330345988274, -0.03032759763300419, -0.014614813961088657, -0.03611212223768234, 0.11501059681177139, 0.11433877795934677, 0.019696475937962532, -0.04344169795513153, -0.0045341430231928825, 0.009732094593346119, -0.06597685813903809, -0.09892185777425766, 0.028556138277053833, 0.028872111812233925, 0.006486900616437197, -0.030460122972726822, 0.1074148565530777, 0.0028660662937909365, -0.1681223064661026, 0.09803245216608047, -0.1454978734254837, -0.13893575966358185, -0.014924702234566212, 0.1282089203596115, -0.028916120529174805, 0.029273254796862602, -0.023388100787997246, -0.012825840152800083, 0.07949179410934448, -0.010074282996356487, -0.06695133447647095, -0.09717464447021484, 0.01728758029639721, -0.07275106757879257, 0.27223652601242065, 0.0025674165226519108, 0.05072332173585892, 0.09320884197950363, 0.00632970267906785, -0.13864335417747498, 0.056013986468315125, 0.04446937516331673, -0.03569450601935387, 0.033913664519786835, 0.17986749112606049, -0.072158083319664, 0.09748772531747818, 0.03297683224081993, -0.17827287316322327, -0.020547686144709587, -0.025412948802113533, -0.005707065109163523, -0.06872113794088364, 0.004436854273080826, -0.08489430695772171, 0.1442941278219223, 0.18185235559940338, -0.007953433319926262, 0.027406441047787666, -0.07448111474514008, 0.031342849135398865, 0.047108203172683716, 0.10373453050851822, 0.002973046386614442, -0.16661332547664642, 0.007796105928719044, -0.07267481833696365, 0.03406558558344841, -0.22716054320335388, -0.11651723086833954, 0.005104650743305683, -0.05659619718790054, -0.05765903368592262, 0.09911203384399414, 0.11035116016864777, 0.02677770145237446, -0.041546907275915146, -0.13469578325748444, -0.03968270868062973, 0.14741691946983337, -0.15685948729515076, -0.04115477204322815 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xmlRoberta_GenData_Double This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results ### Framework versions - Transformers 4.37.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.1
{"license": "mit", "tags": ["generated_from_trainer"], "base_model": "xlm-roberta-base", "model-index": [{"name": "xmlRoberta_GenData_Double", "results": []}]}
text-classification
Constien/xmlRoberta_GenData_Double
[ "transformers", "tensorboard", "safetensors", "xlm-roberta", "text-classification", "generated_from_trainer", "base_model:xlm-roberta-base", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-10T14:04:54+00:00
[]
[]
TAGS #transformers #tensorboard #safetensors #xlm-roberta #text-classification #generated_from_trainer #base_model-xlm-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us
# xmlRoberta_GenData_Double This model is a fine-tuned version of xlm-roberta-base on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results ### Framework versions - Transformers 4.37.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.1
[ "# xmlRoberta_GenData_Double\n\nThis model is a fine-tuned version of xlm-roberta-base on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3", "### Training results", "### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #tensorboard #safetensors #xlm-roberta #text-classification #generated_from_trainer #base_model-xlm-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n", "# xmlRoberta_GenData_Double\n\nThis model is a fine-tuned version of xlm-roberta-base on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3", "### Training results", "### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1" ]
[ 69, 37, 6, 12, 8, 3, 90, 4, 33 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #xlm-roberta #text-classification #generated_from_trainer #base_model-xlm-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n# xmlRoberta_GenData_Double\n\nThis model is a fine-tuned version of xlm-roberta-base on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 5e-05\n- train_batch_size: 16\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3### Training results### Framework versions\n\n- Transformers 4.37.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1" ]
[ -0.08356776088476181, 0.1131780669093132, -0.0020424595568329096, 0.08151137828826904, 0.16415734589099884, 0.01782882772386074, 0.1428142786026001, 0.09786687791347504, -0.09368359297513962, 0.061533767729997635, 0.07376500964164734, 0.08760057389736176, 0.028557458892464638, 0.1409381926059723, -0.06022021919488907, -0.23091116547584534, 0.007577919401228428, -0.005989332217723131, -0.07563283294439316, 0.08642949908971786, 0.102726049721241, -0.11438269913196564, 0.06429262459278107, -0.01667041890323162, -0.170783132314682, 0.03470976650714874, -0.008885747753083706, -0.049410171806812286, 0.08841309696435928, 0.027068687602877617, 0.1174941286444664, 0.01956597901880741, 0.1408947855234146, -0.2102736234664917, 0.008117482997477055, 0.07691167294979095, 0.019248250871896744, 0.0637902170419693, 0.07119226455688477, -0.014638449996709824, 0.09534823149442673, -0.15191756188869476, 0.10095974802970886, 0.019024230539798737, -0.06984581798315048, -0.12276501953601837, -0.08032055199146271, 0.08331095427274704, 0.12188401818275452, 0.08494196087121964, -0.0011123004369437695, 0.14096681773662567, -0.10139140486717224, 0.09146354347467422, 0.16578978300094604, -0.25121891498565674, -0.08525574952363968, 0.09314915537834167, 0.06863132864236832, 0.10433772951364517, -0.1022627204656601, 0.00194830191321671, 0.06368978321552277, 0.022062018513679504, 0.09533783793449402, -0.02242865227162838, -0.06149857118725777, 0.008050095289945602, -0.14634206891059875, 0.005951709114015102, 0.16702350974082947, 0.05302134156227112, -0.05209571123123169, -0.08468539267778397, -0.03537101298570633, -0.06385502219200134, -0.01786835677921772, -0.05938056856393814, 0.045371703803539276, -0.05850839242339134, -0.06001093238592148, -0.058376844972372055, -0.08019191771745682, -0.06179063022136688, 0.008982647210359573, 0.12789656221866608, 0.04865632578730583, -0.003860949072986841, -0.029055310413241386, 0.10568132996559143, 0.01492085587233305, -0.09307906031608582, -0.01141498051583767, 0.01136133074760437, -0.07537968456745148, -0.0794946476817131, -0.034458551555871964, -0.06399212032556534, 0.011914393864572048, 0.1366569548845291, -0.04648159444332123, 0.05719355493783951, 0.015449020080268383, 0.018165459856390953, -0.04052027314901352, 0.135818213224411, -0.03822072222828865, -0.057779181748628616, 0.017685526981949806, 0.0953994020819664, 0.016815047711133957, -0.003755054669454694, -0.10078052431344986, 0.002624400658532977, 0.09338153153657913, 0.07634004950523376, -0.05194198340177536, 0.03292626887559891, -0.021298304200172424, -0.027479657903313637, -0.010055539198219776, -0.12874053418636322, 0.055124860256910324, -0.003224676474928856, -0.06858012080192566, -0.0224835816770792, 0.0330469124019146, 0.007057542912662029, -0.01887248083949089, 0.10522012412548065, -0.08013593405485153, 0.017989883199334145, -0.10625743865966797, -0.09064909815788269, 0.013735783286392689, -0.07007607817649841, -0.007031818386167288, -0.07861550897359848, -0.20747877657413483, -0.05555693805217743, 0.04521593078970909, -0.026568928733468056, -0.03841662406921387, -0.06560125201940536, -0.06030108407139778, -0.00892647635191679, 0.0008603956084698439, 0.10633259266614914, -0.040387146174907684, 0.06846852600574493, 0.003036052919924259, 0.03282869979739189, 0.016331130638718605, 0.03147627413272858, -0.07574629783630371, 0.027146536856889725, -0.13498885929584503, 0.05762972682714462, -0.07846465706825256, 0.04028354957699776, -0.07996963709592819, -0.08917581290006638, 0.00004169300154899247, 0.008519242517650127, 0.028481245040893555, 0.10164345800876617, -0.1593942493200302, -0.03322843462228775, 0.16339941322803497, -0.08173738420009613, -0.06913234293460846, 0.10924386233091354, -0.058049269020557404, 0.04573214799165726, 0.07230688631534576, 0.14593647420406342, 0.08337920904159546, -0.13145564496517181, 0.00614426052197814, -0.011522523127496243, 0.04216257482767105, 0.00870895478874445, 0.056060608476400375, 0.0029863386880606413, 0.013766228221356869, 0.006826731376349926, -0.09400828927755356, -0.003711220808327198, -0.08748937398195267, -0.09163936227560043, -0.044923968613147736, -0.09350205957889557, 0.03630511462688446, 0.027278685942292213, 0.046125248074531555, -0.07958084344863892, -0.09504888206720352, 0.132102370262146, 0.1217183768749237, -0.06764716655015945, -0.003892631968483329, -0.0697658434510231, 0.025151865556836128, -0.05074925348162651, -0.02748699299991131, -0.17231649160385132, -0.11790237575769424, 0.014802614226937294, -0.05886430665850639, 0.05413151532411575, 0.03653837740421295, 0.05362516641616821, 0.09287571161985397, -0.04887852817773819, -0.025027817115187645, -0.06475817412137985, 0.008787659928202629, -0.10594498366117477, -0.20615935325622559, -0.0385371558368206, -0.03884311020374298, 0.17350652813911438, -0.29568448662757874, 0.04501669108867645, -0.07155869901180267, 0.10828312486410141, 0.03947167098522186, -0.03965069353580475, -0.01638967916369438, 0.07615280151367188, -0.025919154286384583, -0.08756545931100845, 0.04013112932443619, -0.009439164772629738, -0.07799477130174637, -0.071637824177742, -0.15338023006916046, 0.09695160388946533, 0.08336151391267776, 0.03085777536034584, -0.11463072150945663, -0.002444129902869463, -0.0486653633415699, -0.04391752555966377, -0.09927948564291, 0.015362770296633244, 0.1499122828245163, -0.01833561435341835, 0.1316504031419754, -0.05345594137907028, -0.04968450590968132, -0.007452556863427162, -0.029803315177559853, -0.009446972981095314, 0.09109839051961899, 0.1105906069278717, -0.1273665428161621, 0.10936659574508667, 0.08132778108119965, -0.10749077051877975, 0.1582215279340744, -0.03661251813173294, -0.06066460907459259, -0.030521439388394356, 0.03237183019518852, 0.00717887980863452, 0.11482167989015579, -0.08170861005783081, -0.004734068177640438, 0.0005850004963576794, 0.0010268486803397536, 0.03907344117760658, -0.1880786269903183, -0.018811754882335663, 0.027167972177267075, -0.04479163885116577, 0.0064894999377429485, -0.035669371485710144, 0.018947452306747437, 0.09220710396766663, -0.0048784781247377396, -0.030010130256414413, 0.0307321697473526, 0.0006239961367100477, -0.10579456388950348, 0.1948239803314209, -0.09283727407455444, -0.13452373445034027, -0.13064992427825928, 0.05546969547867775, -0.06860199570655823, -0.01760414056479931, 0.018932081758975983, -0.08029888570308685, -0.04924497753381729, -0.0986718088388443, -0.03525031730532646, -0.02986116334795952, -0.025700749829411507, 0.018814872950315475, 0.02826048620045185, 0.07133444398641586, -0.12658260762691498, 0.006241294555366039, -0.027447374537587166, -0.10635827481746674, 0.0094221830368042, 0.054799169301986694, 0.12241771817207336, 0.1374320387840271, -0.03173986077308655, 0.005414151586592197, -0.027223963290452957, 0.17764179408550262, -0.0636623352766037, -0.006076580844819546, 0.10927876830101013, 0.0019504039082676172, 0.041664622724056244, 0.11199374496936798, 0.03787357360124588, -0.09216030687093735, 0.02700350061058998, 0.07320088148117065, -0.021260283887386322, -0.23853081464767456, -0.0410555861890316, -0.03371269628405571, -0.05821700021624565, 0.0723189115524292, 0.05095236003398895, 0.04718988016247749, 0.04669661074876785, -0.008568082004785538, 0.05069151520729065, -0.010484568774700165, 0.08585845679044724, 0.058650653809309006, 0.03769616037607193, 0.11154250800609589, -0.04004117101430893, -0.049219317734241486, 0.04400164633989334, -0.0015604221262037754, 0.28734803199768066, 0.0125469621270895, 0.05281548202037811, 0.05721954256296158, 0.12622594833374023, -0.011763317510485649, 0.02864205837249756, 0.034547559916973114, -0.026675445958971977, 0.013086485676467419, -0.06090415641665459, 0.002376721939072013, 0.04200569912791252, -0.03270282223820686, 0.05443922057747841, -0.09625426679849625, 0.03565844148397446, 0.03609207272529602, 0.22042517364025116, 0.0021608155220746994, -0.2978864014148712, -0.06754346191883087, 0.025011558085680008, -0.030024196952581406, -0.035579897463321686, -0.004490859340876341, 0.1270967572927475, -0.13085107505321503, 0.046858061105012894, -0.06855261325836182, 0.0889701172709465, -0.02374354936182499, -0.006546926684677601, 0.02005387470126152, 0.1418067067861557, -0.015032332390546799, 0.08474035561084747, -0.2347087264060974, 0.21994324028491974, 0.01687494106590748, 0.14644679427146912, -0.03892461583018303, 0.02602282166481018, 0.034544575959444046, 0.11140438914299011, 0.09323623776435852, -0.012250890024006367, -0.07065034657716751, -0.1957605928182602, -0.04689832404255867, 0.04240213707089424, 0.11217136681079865, -0.016892554238438606, 0.101634182035923, -0.033812373876571655, 0.004837111569941044, 0.06045934557914734, -0.07381589710712433, -0.18583188951015472, -0.1144431009888649, -0.008008953183889389, 0.00601175194606185, -0.040170419961214066, -0.08663173764944077, -0.10132864117622375, -0.06253158301115036, 0.19316020607948303, -0.03687401860952377, -0.02286044880747795, -0.13014477491378784, 0.07299570739269257, 0.06663554906845093, -0.061666809022426605, 0.009243790060281754, 0.027747003361582756, 0.11667220294475555, 0.022885290905833244, -0.0656125620007515, 0.049251738935709, -0.0890621691942215, -0.15093041956424713, -0.063491590321064, 0.09078925848007202, 0.05804864689707756, 0.0405600443482399, 0.00837974064052105, 0.007527242414653301, 0.027427690103650093, -0.09129223227500916, 0.012209045700728893, 0.09818737208843231, 0.06503631174564362, 0.07741722464561462, -0.11450342833995819, -0.046024031937122345, -0.052323851734399796, -0.03762517496943474, 0.11659611761569977, 0.21234259009361267, -0.0803796797990799, 0.04743761941790581, 0.0696210041642189, -0.11042406409978867, -0.18443767726421356, 0.09671954810619354, 0.08834772557020187, 0.02650582790374756, 0.06785804033279419, -0.1703464537858963, 0.13922452926635742, 0.13133877515792847, -0.015099732205271721, 0.07851413637399673, -0.32512083649635315, -0.1364913135766983, 0.0759241133928299, 0.14400221407413483, 0.023321205750107765, -0.14806951582431793, -0.028330225497484207, -0.024448534473776817, -0.09298660606145859, 0.09980666637420654, -0.12187043577432632, 0.10952800512313843, 0.010659907944500446, 0.06190252676606178, 0.014685727655887604, -0.04182156175374985, 0.1453857719898224, 0.015760494396090508, 0.11636877059936523, -0.05181295424699783, -0.014146567322313786, 0.0792534127831459, -0.06304191797971725, 0.057452745735645294, -0.07146352529525757, 0.04411754384636879, -0.1064453125, -0.038521140813827515, -0.054542068392038345, 0.08186044543981552, -0.054511070251464844, -0.06039433181285858, -0.04238221049308777, 0.05237207189202309, 0.044951654970645905, -0.02435636706650257, 0.05567197501659393, 0.005139345768839121, 0.13325892388820648, 0.0953461080789566, 0.10244482010602951, -0.010579448193311691, -0.07820677012205124, -0.02033723145723343, -0.024309460073709488, 0.06563706696033478, -0.0943794697523117, 0.01596028544008732, 0.11580019444227219, 0.03624460846185684, 0.17269527912139893, 0.04549853131175041, -0.03929496556520462, 0.02167661301791668, 0.0464445985853672, -0.09626892954111099, -0.16143876314163208, -0.002007355447858572, -0.08512690663337708, -0.114984430372715, 0.03383616358041763, 0.09789186716079712, -0.07637261599302292, -0.002604912733659148, -0.02795540541410446, 0.023997079581022263, -0.031042177230119705, 0.18781425058841705, 0.053588949143886566, 0.05700289085507393, -0.0655323714017868, 0.09575655311346054, 0.07728113234043121, -0.07180469483137131, 0.041417498141527176, 0.06985336542129517, -0.07448315620422363, -0.03093857690691948, 0.08081196993589401, 0.22486215829849243, -0.045326270163059235, -0.05524512007832527, -0.12036868929862976, -0.10716541856527328, 0.028387049213051796, 0.13225838541984558, 0.05148785188794136, -0.03356366604566574, -0.024831408634781837, 0.040002334862947464, -0.15719836950302124, 0.10340370982885361, 0.06464707106351852, 0.08289224654436111, -0.1646863967180252, 0.17522472143173218, 0.017576178535819054, 0.014740413054823875, -0.013159327208995819, 0.02558925561606884, -0.10741061717271805, -0.019404364749789238, -0.12402079254388809, -0.005075275897979736, -0.03230720013380051, 0.0070127760991454124, 0.003271323163062334, -0.030742119997739792, -0.04998398572206497, 0.048790693283081055, -0.06857393682003021, -0.05216307193040848, 0.01995028369128704, 0.06455106288194656, -0.12479150295257568, -0.01663622446358204, 0.0008139076526276767, -0.08042185753583908, 0.03975532576441765, 0.037914372980594635, 0.02263573929667473, 0.02619319222867489, -0.18226933479309082, 0.0169310811907053, 0.04628340154886246, 0.022980382665991783, 0.04864092171192169, -0.08702400326728821, -0.001602248870767653, -0.008790257386863232, 0.059556350111961365, 0.024336572736501694, 0.044609781354665756, -0.1199730709195137, 0.0007613225607201457, -0.07866883277893066, -0.06774870306253433, -0.0631745308637619, 0.05631925165653229, 0.09430068731307983, 0.015190677717328072, 0.18347756564617157, -0.10852012038230896, 0.042309198528528214, -0.20068785548210144, -0.030345652252435684, 0.00004894637822872028, -0.031115582212805748, -0.079679936170578, -0.03081498108804226, 0.06873859465122223, -0.05693429708480835, 0.11713458597660065, 0.032352250069379807, 0.05272723734378815, 0.0355968177318573, -0.06269410252571106, 0.004244871437549591, 0.01924477145075798, 0.1686292141675949, 0.04689442366361618, -0.013727816753089428, 0.04053012654185295, 0.005682528018951416, 0.07162921130657196, 0.004297221079468727, 0.19786463677883148, 0.16569121181964874, -0.08294672518968582, 0.07152440398931503, 0.06548483669757843, -0.10353143513202667, -0.1326015591621399, 0.09741197526454926, -0.02315431460738182, 0.1001633033156395, -0.04056688770651817, 0.13748005032539368, 0.14401543140411377, -0.17325709760189056, 0.029043089598417282, -0.0418141670525074, -0.10293465852737427, -0.11921659857034683, -0.04891619831323624, -0.09567061066627502, -0.09971936047077179, 0.019781500101089478, -0.12167342752218246, 0.02847234159708023, 0.06602323800325394, 0.011768038384616375, -0.0020167359616607428, 0.1505063772201538, -0.018638212233781815, 0.026507483795285225, 0.04561711475253105, -0.0009493093239143491, -0.0061418237164616585, -0.06516767293214798, -0.05638982728123665, 0.032546237111091614, -0.024835556745529175, 0.06486887484788895, -0.055542681366205215, -0.003295092610642314, 0.044079191982746124, -0.023274099454283714, -0.06250910460948944, 0.036162614822387695, 0.018406396731734276, 0.0398167259991169, 0.06202453374862671, 0.04862050712108612, -0.030098460614681244, -0.02966308407485485, 0.252339631319046, -0.06418249756097794, -0.08935225754976273, -0.11775318533182144, 0.22161541879177094, 0.045231644064188004, 0.0017135581001639366, 0.05008229985833168, -0.11127695441246033, 0.0313834547996521, 0.1660824865102768, 0.1684858649969101, -0.048311375081539154, -0.008716325275599957, -0.028298307210206985, -0.023608826100826263, -0.05755394324660301, 0.12261199951171875, 0.08143693953752518, 0.07659527659416199, -0.04007285088300705, -0.0011599463177844882, -0.0130050303414464, -0.024545229971408844, -0.10773826390504837, 0.04305059835314751, 0.03306875750422478, 0.0054744938388466835, -0.020490074530243874, 0.0814383253455162, 0.0027225192170590162, -0.16111785173416138, 0.07159152626991272, -0.13629798591136932, -0.164171501994133, -0.03219759836792946, 0.0946953296661377, -0.03262479230761528, 0.04769578576087952, -0.026499293744564056, -0.02442566119134426, 0.10904210805892944, -0.019568312913179398, -0.04971575736999512, -0.08702147752046585, 0.06248052790760994, -0.06368279457092285, 0.2268555760383606, -0.009990588761866093, 0.07094763219356537, 0.10885593295097351, 0.006272402126342058, -0.09729922562837601, 0.08458259701728821, 0.046594392508268356, -0.05397757142782211, 0.07023145258426666, 0.13868942856788635, -0.06268349289894104, 0.10679212957620621, 0.041391100734472275, -0.11490187048912048, 0.0056647746823728085, -0.057452503591775894, -0.06387865543365479, -0.07896646857261658, 0.013307360000908375, -0.07343804091215134, 0.1455385386943817, 0.20290015637874603, -0.02712336555123329, 0.024820812046527863, -0.0667681097984314, 0.04218606650829315, 0.06901448220014572, 0.09944040328264236, -0.022530900314450264, -0.2187507003545761, 0.03568357601761818, -0.02162090688943863, 0.02401922456920147, -0.28299063444137573, -0.09141562879085541, 0.013100167736411095, -0.05737555772066116, -0.06746021658182144, 0.08906529098749161, 0.13275150954723358, 0.04059172794222832, -0.044880613684654236, -0.15321403741836548, -0.05717765912413597, 0.15241345763206482, -0.1264445185661316, -0.05468975380063057 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-large-squad-model2 This model is a fine-tuned version of [t5-large](https://huggingface.co/t5-large) on the squad dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 92 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.1+cu121 - Datasets 2.15.0 - Tokenizers 0.15.0
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["varun-v-rao/squad"], "base_model": "t5-large", "model-index": [{"name": "t5-large-squad-model2", "results": []}]}
question-answering
varun-v-rao/t5-large-squad-model2
[ "transformers", "tensorboard", "safetensors", "t5", "question-answering", "generated_from_trainer", "dataset:varun-v-rao/squad", "base_model:t5-large", "license:apache-2.0", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T14:06:04+00:00
[]
[]
TAGS #transformers #tensorboard #safetensors #t5 #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-t5-large #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us
# t5-large-squad-model2 This model is a fine-tuned version of t5-large on the squad dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 92 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.1+cu121 - Datasets 2.15.0 - Tokenizers 0.15.0
[ "# t5-large-squad-model2\n\nThis model is a fine-tuned version of t5-large on the squad dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 92\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3", "### Training results", "### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0" ]
[ "TAGS\n#transformers #tensorboard #safetensors #t5 #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-t5-large #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us \n", "# t5-large-squad-model2\n\nThis model is a fine-tuned version of t5-large on the squad dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 92\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3", "### Training results", "### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0" ]
[ 81, 32, 6, 12, 8, 3, 90, 4, 33 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #t5 #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-t5-large #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us \n# t5-large-squad-model2\n\nThis model is a fine-tuned version of t5-large on the squad dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 92\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3### Training results### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0" ]
[ -0.0839281976222992, 0.1261083334684372, -0.003062264062464237, 0.09963858872652054, 0.13135892152786255, 0.012065015733242035, 0.11763228476047516, 0.1431812047958374, -0.10027769953012466, 0.05847493186593056, 0.06298580765724182, 0.058026451617479324, 0.044224463403224945, 0.12045852094888687, -0.034110307693481445, -0.20944428443908691, 0.008230980485677719, -0.024585377424955368, -0.07987969368696213, 0.11169750243425369, 0.10547711700201035, -0.08776256442070007, 0.09019466489553452, -0.01803523488342762, -0.13123410940170288, 0.055404581129550934, -0.016156500205397606, -0.04871221259236336, 0.10830388963222504, 0.037360187619924545, 0.07608335465192795, 0.013134647160768509, 0.11841697245836258, -0.23230670392513275, 0.008173879235982895, 0.09913269430398941, 0.009827329777181149, 0.06901858001947403, 0.04457641765475273, 0.010080515407025814, 0.1014266237616539, -0.15918457508087158, 0.09575871378183365, 0.038182880729436874, -0.07762988656759262, -0.15214751660823822, -0.08740975707769394, 0.08058422803878784, 0.09363251179456711, 0.10123783349990845, 0.002513112733140588, 0.16347931325435638, -0.07089421898126602, 0.09687786549329758, 0.21524114906787872, -0.27519577741622925, -0.051773570477962494, 0.06465653330087662, 0.06387705355882645, 0.09176666289567947, -0.10714096575975418, -0.013616802170872688, 0.05330894887447357, 0.023993054404854774, 0.09900592267513275, -0.024224957451224327, -0.12270430475473404, 0.01981201581656933, -0.1330566257238388, -0.026274116709828377, 0.19247335195541382, 0.05531616508960724, -0.03306442126631737, -0.09265746921300888, -0.07251456379890442, -0.07065938413143158, 0.0033086352050304413, -0.055225301533937454, 0.04939652606844902, -0.05036013200879097, -0.04117193818092346, -0.06376560777425766, -0.08855835348367691, -0.07557594031095505, -0.006487980484962463, 0.06795371323823929, 0.060516294091939926, 0.02030317671597004, -0.052223484963178635, 0.09670409560203552, -0.000922380480915308, -0.11623311787843704, -0.02956419251859188, 0.010396474041044712, -0.1006184071302414, -0.05367468670010567, -0.016288749873638153, -0.04984374716877937, 0.026406390592455864, 0.14233064651489258, -0.07116742432117462, 0.04758491367101669, 0.006304616574198008, 0.0020201641600579023, -0.023326370865106583, 0.12988999485969543, -0.0747334212064743, -0.03606707602739334, 0.0195129606872797, 0.09720822423696518, 0.027357831597328186, 0.0006529309903271496, -0.0756712555885315, -0.016676167026162148, 0.10392086952924728, 0.09382186084985733, -0.021162066608667374, 0.04972006753087044, -0.005882317200303078, -0.022401096299290657, 0.028351997956633568, -0.1451730728149414, 0.023124340921640396, -0.014342641457915306, -0.07552029937505722, -0.05094066262245178, 0.04416579753160477, -0.013332963921129704, -0.042830489575862885, 0.041396647691726685, -0.08207105845212936, -0.018468378111720085, -0.06805264204740524, -0.07735759019851685, 0.034603558480739594, -0.05059023201465607, -0.010439747013151646, -0.08485570549964905, -0.2111506313085556, -0.03390905261039734, 0.018963005393743515, -0.04166364297270775, -0.04795591160655022, -0.03588128834962845, -0.08114167302846909, -0.007020397111773491, -0.010196682997047901, 0.10208071023225784, -0.05284076929092407, 0.07554060220718384, 0.0006300450768321753, 0.026895886287093163, 0.028163552284240723, 0.03207937628030777, -0.10759899020195007, 0.027773508802056313, -0.12555663287639618, 0.048252034932374954, -0.0745185911655426, 0.04513080418109894, -0.1317071169614792, -0.0913582369685173, 0.014598852023482323, -0.014930770732462406, 0.03973560035228729, 0.12607920169830322, -0.17827369272708893, -0.01184755191206932, 0.16175490617752075, -0.09103478491306305, -0.12444166094064713, 0.10817787051200867, -0.035492949187755585, 0.031127873808145523, 0.07547077536582947, 0.14809326827526093, 0.11798910051584244, -0.1529514193534851, -0.023454764857888222, 0.00978886429220438, 0.026510832831263542, 0.005811793264001608, 0.06492045521736145, -0.002422938821837306, 0.0324135348200798, 0.0033832855988293886, -0.07774242758750916, 0.004746237304061651, -0.06814143806695938, -0.08468984812498093, -0.06656024605035782, -0.09494784474372864, 0.016896408051252365, 0.04425622522830963, 0.016107995063066483, -0.07891404628753662, -0.12157101929187775, 0.10181677341461182, 0.13314993679523468, -0.05751636251807213, 0.017445892095565796, -0.0826665535569191, 0.04223857820034027, -0.035234954208135605, -0.009665212593972683, -0.17604783177375793, -0.15497589111328125, 0.030343323945999146, -0.07495743036270142, 0.0386577807366848, 0.028882358223199844, 0.06667955964803696, 0.06197921559214592, -0.06662274152040482, -0.03190949186682701, -0.0694076269865036, 0.0028224631678313017, -0.08084381371736526, -0.19966425001621246, -0.03640775382518768, -0.024320878088474274, 0.11171878129243851, -0.24468429386615753, 0.03274338319897652, -0.0032370334956794977, 0.12300720810890198, 0.03792393207550049, -0.03558218851685524, 0.010310020297765732, 0.015933429822325706, -0.021904312074184418, -0.09577424079179764, 0.02584630250930786, -0.013548243790864944, -0.07344982028007507, -0.03017893061041832, -0.14608134329319, 0.09459366649389267, 0.0761425793170929, 0.0656321793794632, -0.09726715832948685, 0.011707599274814129, -0.06335706263780594, -0.05005437880754471, -0.08289062976837158, -0.028516694903373718, 0.13600964844226837, 0.006424510385841131, 0.11603967845439911, -0.07780507951974869, -0.07477131485939026, -0.0017668004147708416, 0.0018605522345751524, -0.0017423651879653335, 0.09412875771522522, 0.06281629204750061, -0.1333269476890564, 0.10641986131668091, 0.10460934042930603, -0.043322429060935974, 0.13723532855510712, -0.06842523068189621, -0.09290395677089691, -0.02536940760910511, 0.04502733424305916, -0.004199197981506586, 0.12755340337753296, -0.0779266282916069, 0.009713784791529179, 0.02394826151430607, 0.007415453903377056, 0.02141011320054531, -0.1676000952720642, -0.02255023829638958, 0.02750781551003456, -0.06232931464910507, -0.0027935290709137917, -0.008672268129885197, 0.028863131999969482, 0.09656790643930435, -0.0019535499159246683, -0.03642662242054939, 0.020969947800040245, -0.01372967567294836, -0.09690739214420319, 0.1940627098083496, -0.09138940274715424, -0.16768701374530792, -0.11787500977516174, 0.06430284678936005, -0.058576930314302444, -0.03869358077645302, 0.027305535972118378, -0.08454590290784836, -0.05970468372106552, -0.12390334159135818, 0.01670565828680992, -0.004053461365401745, -0.01708490401506424, -0.00236968114040792, 0.037480976432561874, 0.09131629765033722, -0.1461399644613266, 0.023405805230140686, -0.001543866004794836, -0.12375885993242264, -0.02270522341132164, 0.0442158579826355, 0.12361281365156174, 0.12604370713233948, -0.017666274681687355, 0.01435889769345522, -0.03932787850499153, 0.19681385159492493, -0.07420585304498672, 0.027216849848628044, 0.13123352825641632, 0.011457717046141624, 0.0512230321764946, 0.14328128099441528, 0.02061961032450199, -0.08868986368179321, 0.04612944647669792, 0.08952318876981735, -0.017557840794324875, -0.2710972726345062, -0.027723759412765503, -0.02768908068537712, -0.025134418159723282, 0.07308977842330933, 0.07395222038030624, 0.040232814848423004, 0.030448652803897858, -0.019858328625559807, 0.023309430107474327, 0.007207927759736776, 0.07565172016620636, 0.0919472873210907, 0.029040170833468437, 0.08902935683727264, -0.05398954078555107, -0.032996274530887604, 0.06882938742637634, 0.03148900717496872, 0.2680175304412842, -0.029900986701250076, 0.1270894855260849, 0.048716574907302856, 0.1359742134809494, -0.038751788437366486, 0.03845788538455963, 0.005879363976418972, 0.003590828739106655, 0.01499318890273571, -0.06965764611959457, 0.015490095131099224, 0.04385251924395561, -0.04226115718483925, 0.05273814499378204, -0.07301559299230576, 0.05680014193058014, 0.0428449809551239, 0.258347749710083, 0.04114618897438049, -0.2807044982910156, -0.06547261029481888, 0.01581575907766819, -0.04947369918227196, -0.041178084909915924, 0.03558659553527832, 0.15338380634784698, -0.10331800580024719, 0.05324047803878784, -0.052818089723587036, 0.08227041363716125, -0.0332508310675621, -0.006562414579093456, 0.06008775904774666, 0.10793577879667282, -0.01466459408402443, 0.09867450594902039, -0.2054762840270996, 0.2212868481874466, 0.019199585542082787, 0.09849648922681808, -0.0537620410323143, 0.027258088812232018, 0.01388336531817913, 0.10601738095283508, 0.1394767165184021, -0.006770754233002663, -0.03920384496450424, -0.14102481305599213, -0.08395077288150787, 0.044270116835832596, 0.09617991000413895, -0.02331448718905449, 0.09025581181049347, -0.05987977609038353, -0.0045128632336854935, 0.0571756474673748, -0.0553128719329834, -0.16500459611415863, -0.11064475029706955, 0.008484463207423687, 0.0011187626514583826, -0.05869697034358978, -0.09411495923995972, -0.09527410566806793, -0.03196074813604355, 0.1659775823354721, -0.003866406623274088, -0.05522727966308594, -0.12777341902256012, 0.052108969539403915, 0.11907994002103806, -0.07553528249263763, 0.010025964118540287, 0.021394789218902588, 0.10863605886697769, 0.045611847192049026, -0.09169382601976395, 0.06781083345413208, -0.06395947188138962, -0.16807101666927338, -0.04389529302716255, 0.13599520921707153, 0.03792949020862579, 0.03889857232570648, -0.0003481656312942505, 0.008316682651638985, 0.026777684688568115, -0.08510306477546692, -0.0014019650407135487, 0.06908272206783295, 0.06039856746792793, 0.055292949080467224, -0.08211416751146317, -0.010816353373229504, -0.0460919551551342, -0.028979342430830002, 0.12151381373405457, 0.19808121025562286, -0.07664288580417633, 0.07020253688097, 0.07568889111280441, -0.0810757502913475, -0.18813800811767578, 0.057476989924907684, 0.04879027232527733, -0.003441792679950595, 0.07332808524370193, -0.1549539864063263, 0.11425995826721191, 0.10453013330698013, -0.018607595935463905, 0.0703471228480339, -0.3515491485595703, -0.14187200367450714, 0.08628997951745987, 0.1229277178645134, -0.016710126772522926, -0.16788429021835327, -0.036095306277275085, -0.019366707652807236, -0.12673474848270416, 0.10617314279079437, -0.1308029145002365, 0.0904388353228569, -0.006109047215431929, 0.0796297937631607, 0.024943538010120392, -0.03742226958274841, 0.11749859154224396, 0.02955463156104088, 0.09640388935804367, -0.06620027124881744, 0.00005063251592218876, 0.11677919328212738, -0.07344992458820343, 0.08932027965784073, -0.0741121917963028, 0.08342885971069336, -0.12626494467258453, -0.020201807841658592, -0.07732761651277542, 0.06287550181150436, -0.060852810740470886, -0.06261437386274338, -0.06867826730012894, 0.06399078667163849, 0.06451881676912308, -0.035862572491168976, 0.09243885427713394, 0.018544329330325127, 0.09275095164775848, 0.1065312847495079, 0.1063532754778862, 0.015045384876430035, -0.11864463239908218, 0.000017849089999799617, -0.020689193159341812, 0.055316224694252014, -0.1493052989244461, 0.04043823853135109, 0.1179082840681076, 0.045591894537210464, 0.13546763360500336, 0.025959674268960953, -0.06078602373600006, -0.0209849551320076, 0.03881833702325821, -0.12147563695907593, -0.20906564593315125, -0.012709543108940125, -0.04548925906419754, -0.14717940986156464, 0.04779060557484627, 0.10240687429904938, -0.07102843374013901, -0.004510396160185337, -0.011736433953046799, 0.04795883968472481, -0.01813504472374916, 0.17448288202285767, 0.06531528383493423, 0.06371259689331055, -0.07295311242341995, 0.13005128502845764, 0.07830176502466202, -0.0668882504105568, 0.04375062510371208, 0.08446737378835678, -0.07878647744655609, -0.03125622868537903, 0.07505245506763458, 0.15181714296340942, -0.04045454040169716, -0.058622319251298904, -0.09682145714759827, -0.08634954690933228, 0.04188886657357216, 0.15657295286655426, 0.04036259651184082, 0.004686446860432625, -0.0021561039611697197, 0.02002524584531784, -0.12902860343456268, 0.12812539935112, 0.04141845554113388, 0.0700337216258049, -0.16431033611297607, 0.0889073833823204, 0.012128245085477829, 0.043359238654375076, -0.017898913472890854, 0.040353402495384216, -0.09532704204320908, -0.01772008277475834, -0.1447451263666153, -0.003086257493123412, -0.01317469123750925, 0.014665401540696621, -0.005853652022778988, -0.05575815588235855, -0.048601459711790085, 0.06140086054801941, -0.05961346626281738, -0.05857682600617409, 0.026708418503403664, 0.07027518004179001, -0.1751018762588501, -0.025129331275820732, 0.024960173293948174, -0.0895155742764473, 0.08496726304292679, 0.019803646951913834, 0.00711607001721859, 0.04133259877562523, -0.11655022203922272, 0.0017125135054811835, 0.025988224893808365, 0.03829248622059822, 0.05650504678487778, -0.10616175085306168, -0.006648528855293989, -0.024453047662973404, 0.026401469483971596, 0.021597834303975105, 0.026651309803128242, -0.11450669914484024, -0.004813461098819971, -0.06269919872283936, -0.04914041981101036, -0.053476933389902115, 0.03652093932032585, 0.08632200211286545, 0.014656415209174156, 0.16445760428905487, -0.07905375212430954, 0.037980955094099045, -0.22045069932937622, -0.026244426146149635, 0.01592385023832321, -0.03226268291473389, -0.058460526168346405, -0.013634145259857178, 0.06988880783319473, -0.07099566608667374, 0.11029772460460663, -0.006629397161304951, 0.08949234336614609, 0.054430242627859116, -0.02328120358288288, 0.0001429155090590939, 0.012257945723831654, 0.16077645123004913, 0.01872006244957447, -0.017407819628715515, 0.06745051592588425, -0.0414283461868763, 0.060396403074264526, -0.002627432346343994, 0.19473440945148468, 0.1560627818107605, -0.04781947657465935, 0.055715810507535934, 0.08512163907289505, -0.10874835401773453, -0.12130145728588104, 0.076547771692276, -0.01866471767425537, 0.11095082014799118, -0.04885119944810867, 0.16227298974990845, 0.1464015692472458, -0.15504398941993713, 0.04645893722772598, -0.06448610126972198, -0.10557472705841064, -0.11600614339113235, -0.056020770221948624, -0.086344875395298, -0.11835401505231857, 0.020476821810007095, -0.1264120638370514, 0.031245436519384384, 0.07587498426437378, 0.017492186278104782, -0.004459212068468332, 0.1782204955816269, -0.013752544298768044, 0.01731204241514206, 0.03664284199476242, 0.01656637154519558, -0.0048347944393754005, -0.04331904649734497, -0.02818288654088974, 0.05729083716869354, 0.0020359286572784185, 0.048150379210710526, -0.03658107668161392, 0.01167744118720293, 0.04917609691619873, -0.030710870400071144, -0.0569467656314373, 0.01393153052777052, 0.019030265510082245, 0.02428223192691803, 0.046454738825559616, 0.0692119374871254, 0.0031428951770067215, -0.02897767163813114, 0.2809711992740631, -0.07201123982667923, -0.0987803041934967, -0.1269555240869522, 0.17352034151554108, 0.027358166873455048, -0.011293497867882252, 0.06895182281732559, -0.13112309575080872, -0.008546622470021248, 0.17493797838687897, 0.14957621693611145, -0.040105726569890976, -0.011784175410866737, -0.029288465157151222, -0.009736468084156513, -0.0567866787314415, 0.10474211722612381, 0.10415248572826385, 0.059311386197805405, -0.04549558833241463, -0.016818277537822723, -0.0015276905614882708, -0.020286427810788155, -0.0667034238576889, 0.08278689533472061, 0.021286407485604286, 0.019790183752775192, -0.023807983845472336, 0.08524836599826813, 0.00005774085730081424, -0.19958144426345825, 0.027198106050491333, -0.15020152926445007, -0.1605449765920639, -0.033723656088113785, 0.11400729417800903, -0.030072279274463654, 0.03725128993391991, -0.013776549138128757, -0.004369073547422886, 0.12383858859539032, -0.007117966189980507, -0.09742242842912674, -0.11070932447910309, 0.08408721536397934, -0.08965811878442764, 0.24401839077472687, -0.0065744356252253056, 0.058194879442453384, 0.11118236929178238, -0.02101377584040165, -0.14902886748313904, 0.06021351367235184, 0.07102279365062714, -0.05791439116001129, 0.022585904225707054, 0.14585039019584656, -0.036786455661058426, 0.10659541189670563, 0.04141741245985031, -0.08615081012248993, -0.02117605321109295, -0.014005827717483044, -0.021288903430104256, -0.1071854904294014, 0.007492178585380316, -0.07142889499664307, 0.14643456041812897, 0.19126097857952118, -0.04152262583374977, 0.03364183381199837, -0.08757299929857254, 0.024191370233893394, 0.06556772440671921, 0.05851958692073822, 0.010666498914361, -0.17674310505390167, 0.036180321127176285, -0.008579286746680737, 0.03538830578327179, -0.24115124344825745, -0.08693405985832214, 0.049991462379693985, -0.04674219712615013, -0.07683096081018448, 0.11560467630624771, 0.11185789108276367, 0.04734513536095619, -0.03202183172106743, -0.11555701494216919, -0.04850735887885094, 0.14752691984176636, -0.16545924544334412, -0.05766911059617996 ]
null
null
stable-baselines3
# **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
{"library_name": "stable-baselines3", "tags": ["LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "PPO", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "LunarLander-v2", "type": "LunarLander-v2"}, "metrics": [{"type": "mean_reward", "value": "265.96 +/- 21.30", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
arekpaterak/ppo-LunarLander-v2
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
2024-02-10T14:06:58+00:00
[]
[]
TAGS #stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
# PPO Agent playing LunarLander-v2 This is a trained model of a PPO agent playing LunarLander-v2 using the stable-baselines3 library. ## Usage (with Stable-baselines3) TODO: Add your code
[ "# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.", "## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ "TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n", "# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.", "## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ 39, 41, 17 ]
[ "passage: TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ 0.03942384943366051, 0.04900386184453964, -0.005304091144353151, 0.026427261531352997, 0.107408307492733, -0.026511888951063156, 0.11188238859176636, 0.0814051404595375, 0.10722193866968155, 0.04762078449130058, 0.08338645845651627, 0.06030960753560066, 0.05080918222665787, 0.2571701407432556, 0.04754156619310379, -0.22987541556358337, 0.036159250885248184, -0.04869936779141426, 0.12395193427801132, 0.07178173214197159, -0.0038484656251966953, -0.06485428661108017, 0.020415637642145157, -0.013290755450725555, 0.05367108806967735, 0.04282612353563309, -0.01716216839849949, -0.08207534998655319, 0.07169748842716217, -0.06345846503973007, 0.06986866891384125, 0.07677983492612839, 0.13218913972377777, -0.17832116782665253, 0.029566360637545586, 0.02571309357881546, -0.07189024239778519, 0.01342033501714468, 0.008019951172173023, 0.05120139941573143, 0.17303818464279175, 0.019879888743162155, 0.07844575494527817, -0.0025605305563658476, -0.15412317216396332, -0.018950799480080605, 0.0436202734708786, 0.12546207010746002, 0.08808347582817078, 0.04605821147561073, 0.01970590092241764, 0.17503218352794647, -0.054352790117263794, -0.028833400458097458, 0.21759237349033356, -0.2881564497947693, -0.031460098922252655, 0.321048766374588, 0.06997483223676682, 0.09725230932235718, -0.07540661096572876, -0.03619609400629997, 0.007783263456076384, -0.013137873262166977, -0.028666524216532707, -0.07447073608636856, 0.17313385009765625, 0.05152064561843872, -0.05057951435446739, -0.09541505575180054, 0.16948209702968597, 0.006921638268977404, 0.0018855923553928733, -0.019282981753349304, 0.009060598909854889, 0.07402525842189789, -0.016097044572234154, -0.07255112379789352, 0.057438433170318604, 0.05330665782094002, 0.019649166613817215, -0.1435653269290924, -0.10762494057416916, -0.022740179672837257, -0.008012006990611553, 0.17786912620067596, -0.009255532175302505, 0.042902372777462006, 0.003065188182517886, 0.10384012013673782, -0.12480384111404419, -0.03354184702038765, -0.0454259067773819, -0.07565800100564957, -0.0223417766392231, -0.02058211714029312, -0.03580251708626747, 0.07184842973947525, 0.11971849203109741, 0.027368178591132164, 0.09350208193063736, 0.047715865075588226, -0.03206788748502731, 0.06343851238489151, 0.05555703118443489, 0.14222665131092072, 0.05807621404528618, 0.012854371219873428, 0.13179877400398254, 0.055213116109371185, 0.033023182302713394, -0.0613492950797081, -0.18252409994602203, 0.07489913702011108, -0.07031869143247604, 0.007941240444779396, 0.12051256000995636, -0.04480670019984245, -0.1183447614312172, -0.037500523030757904, -0.017392054200172424, -0.06224250793457031, -0.025395862758159637, 0.0547584593296051, -0.02883218228816986, -0.03973718360066414, 0.0011496668448671699, 0.09384800493717194, 0.00953749567270279, -0.1752052903175354, 0.03303423151373863, -0.025042934343218803, -0.10782608389854431, 0.009975161403417587, 0.0022444494534283876, 0.03394931182265282, 0.04408763721585274, -0.11822668462991714, -0.30899152159690857, -0.07652641832828522, 0.05490870401263237, -0.06516939401626587, -0.18425025045871735, -0.13193942606449127, 0.02454492449760437, -0.09037084132432938, -0.044885024428367615, -0.12759265303611755, -0.028549788519740105, 0.01743689924478531, 0.011519349180161953, 0.10758619755506516, -0.0106219332665205, -0.012188062071800232, -0.1571401208639145, 0.008273907005786896, -0.20951123535633087, 0.0890483483672142, -0.019150104373693466, 0.037884220480918884, -0.032381169497966766, -0.07404014468193054, 0.030707746744155884, 0.052499737590551376, -0.01474119070917368, 0.13510210812091827, -0.15592676401138306, -0.03691192343831062, -0.007996266707777977, -0.13611900806427002, -0.04786273464560509, -0.10358831286430359, -0.04357128217816353, 0.13354332745075226, 0.018664736300706863, 0.15356586873531342, -0.08709818124771118, -0.0722038671374321, 0.20489206910133362, -0.010411538183689117, -0.12820468842983246, -0.076752208173275, 0.10165707021951675, 0.021510310471057892, -0.056606587022542953, -0.02523270808160305, -0.1839766949415207, -0.0152357779443264, -0.04550420492887497, -0.047039128839969635, 0.01796751655638218, -0.010888241231441498, 0.13837894797325134, 0.08494598418474197, 0.05018039792776108, -0.06086122244596481, -0.006730288732796907, 0.10779471695423126, 0.08823856711387634, 0.008680110797286034, 0.023406028747558594, -0.05774238705635071, 0.09552932530641556, -0.04003755748271942, -0.0142367510125041, -0.08283266425132751, -0.036246106028556824, -0.026256313547492027, 0.17507147789001465, 0.09440762549638748, 0.2257927656173706, 0.09567736834287643, 0.039160262793302536, 0.031270865350961685, -0.13181598484516144, -0.1425403207540512, -0.0017254541162401438, 0.09020978957414627, -0.14270411431789398, -0.04119925573468208, -0.08974775671958923, -0.17768175899982452, -0.12202505767345428, 0.0006432619411498308, -0.17960017919540405, 0.06390921026468277, 0.05408334732055664, -0.035177867859601974, 0.03272094577550888, 0.13032332062721252, -0.011533179320394993, -0.03967514634132385, 0.0831870287656784, 0.0379033200442791, -0.041234664618968964, -0.021742934361100197, 0.11885567009449005, 0.15673065185546875, 0.13124459981918335, -0.03511447086930275, 0.004914294462651014, 0.07076404243707657, -0.02309088408946991, 0.06539414077997208, 0.0558244064450264, 0.20973342657089233, 0.188301220536232, 0.038996949791908264, 0.008822928182780743, -0.07048165798187256, 0.0855446457862854, -0.0742373839020729, -0.14302679896354675, -0.05579735338687897, 0.08729292452335358, 0.016605578362941742, 0.023469142615795135, 0.08711627870798111, 0.024545932188630104, 0.09132762253284454, 0.15968108177185059, 0.01990218088030815, -0.09659269452095032, -0.050218869000673294, 0.01175848301500082, 0.027713103219866753, 0.04794301092624664, -0.04514073207974434, -0.00937939714640379, 0.017020760104060173, -0.10303554683923721, 0.031789086759090424, -0.1413339376449585, -0.1358717679977417, 0.044326696544885635, 0.003906996920704842, 0.010907664895057678, 0.02786896750330925, -0.0038291432429105043, 0.019039705395698547, 0.04351753741502762, -0.06975466758012772, 0.047416772693395615, -0.024745507165789604, -0.020031947642564774, 0.03340689837932587, -0.057257164269685745, -0.205775648355484, -0.17696654796600342, 0.00013708483311347663, -0.09910997003316879, 0.10194740444421768, 0.018308809027075768, -0.12373185902833939, 0.047737859189510345, -0.05822649225592613, 0.027574289590120316, -0.01875593699514866, -0.049130141735076904, 0.10507171601057053, 0.1525275856256485, -0.016146350651979446, 0.018018173053860664, -0.04865182936191559, -0.10157987475395203, -0.19632206857204437, 0.0691583976149559, 0.04680244252085686, 0.014610917307436466, 0.10669491440057755, 0.018072687089443207, 0.02367905154824257, -0.007674071006476879, -0.016521066427230835, -0.011659215204417706, -0.08781040459871292, 0.31909599900245667, 0.04510033503174782, -0.025173069909214973, 0.02041010931134224, -0.0043001663871109486, -0.028083480894565582, 0.03263787180185318, -0.0985708013176918, -0.07548979669809341, -0.08774089068174362, -0.04367410019040108, -0.09784720093011856, 0.053299110382795334, 0.05916472524404526, 0.003188040340319276, -0.07727594673633575, 0.04221395403146744, 0.11369874328374863, -0.0923808291554451, -0.07137343287467957, 0.07477962225675583, 0.0972946360707283, -0.07331304252147675, 0.00012658814375754446, 0.00874367356300354, 0.023951783776283264, 0.037102166563272476, 0.06778035312891006, -0.03966575115919113, 0.08589404821395874, -0.19917890429496765, 0.0372927263379097, 0.106058269739151, 0.023754918947815895, 0.0638108178973198, 0.07643651217222214, -0.1058402881026268, -0.008500572293996811, -0.032518330961465836, -0.21341575682163239, 0.1668180525302887, 0.1355515867471695, 0.06788124144077301, -0.025637222453951836, -0.00461410591378808, -0.0649740919470787, 0.05773647129535675, 0.02723747305572033, -0.14758841693401337, 0.004883295856416225, 0.06064270809292793, 0.026899009943008423, 0.01614922471344471, 0.07971042394638062, 0.014697225764393806, -0.1801026314496994, -0.014406266622245312, 0.10730406641960144, 0.002390873385593295, 0.0053148469887673855, -0.03175045922398567, -0.1755964607000351, 0.0751047357916832, 0.004285442177206278, 0.07233936339616776, -0.1676585078239441, 0.14297930896282196, -0.10089799761772156, 0.07726949453353882, -0.004285062663257122, -0.021311495453119278, 0.02507244050502777, -0.0541163794696331, 0.15163759887218475, 0.01058570109307766, -0.021810131147503853, -0.1200498715043068, -0.1717042326927185, -0.019227758049964905, -0.11788936704397202, -0.11679866164922714, 0.050424277782440186, 0.062185097485780716, 0.04923136904835701, -0.061147067695856094, 0.1518532931804657, -0.047422297298908234, 0.060713399201631546, -0.06893875449895859, -0.06755045056343079, 0.03764858841896057, -0.12588608264923096, -0.08176055550575256, 0.05573027580976486, 0.19166934490203857, 0.15833087265491486, -0.02816431224346161, -0.03472423925995827, -0.047419581562280655, -0.006212298292666674, -0.007802055217325687, 0.0275666993111372, 0.023223137483000755, 0.07315318286418915, -0.07681374251842499, -0.11649256944656372, 0.033787861466407776, -0.06713802367448807, -0.055589709430933, -0.015439179725944996, 0.1513158082962036, 0.04671623185276985, 0.07720734924077988, -0.018946662545204163, 0.03887668624520302, -0.001724981120787561, -0.056474871933460236, 0.16197094321250916, 0.03885216265916824, -0.05193585529923439, 0.06837689876556396, 0.053174007683992386, 0.043745119124650955, 0.03011113777756691, -0.026783017441630363, 0.206032395362854, 0.1980147808790207, 0.014206883497536182, 0.2175983190536499, 0.03177616000175476, -0.03772832080721855, -0.1300560086965561, -0.065880686044693, -0.006372632458806038, 0.03559038043022156, 0.08070417493581772, -0.18207235634326935, -0.015011128038167953, -0.05689644813537598, -0.034518610686063766, -0.15059494972229004, -0.28553900122642517, -0.05957856774330139, 0.20075850188732147, 0.14706264436244965, 0.27519428730010986, -0.10432573407888412, 0.035197313874959946, 0.02663275972008705, -0.04912831634283066, -0.006501141935586929, 0.00018665487004909664, 0.10268618166446686, -0.15421873331069946, 0.1176437959074974, 0.08486983180046082, -0.019002694636583328, 0.01058861706405878, -0.1619086116552353, 0.00936629343777895, -0.12191236019134521, 0.05354422330856323, 0.1400289237499237, -0.048128653317689896, -0.054873593151569366, 0.14033560454845428, -0.024562934413552284, -0.22685599327087402, -0.04648222774267197, -0.043600670993328094, -0.010640020482242107, 0.026607351377606392, -0.1013401448726654, 0.04101909324526787, 0.1330099105834961, 0.009380043484270573, 0.1147187277674675, 0.11749245226383209, -0.052566803991794586, 0.10792597383260727, 0.2257719188928604, -0.018785694614052773, 0.04689010605216026, -0.12743118405342102, -0.0012336712097749114, -0.028270328417420387, 0.013657891191542149, -0.09504974633455276, -0.09938385337591171, 0.02366873063147068, 0.02872389927506447, 0.009118586778640747, 0.0921793207526207, -0.029922157526016235, 0.0759170651435852, 0.06817561388015747, -0.13014446198940277, -0.16288450360298157, 0.015828335657715797, -0.007344507612287998, 0.08354310691356659, 0.00027861111448146403, 0.08878035843372345, -0.11932205408811569, -0.018093237653374672, -0.03153328225016594, -0.03319635987281799, -0.130486860871315, -0.07138993591070175, 0.06156524643301964, 0.028095467016100883, -0.06602972000837326, 0.1398407518863678, 0.026440169662237167, 0.15942534804344177, 0.049197953194379807, 0.012499804608523846, 0.07227300107479095, -0.05345509201288223, 0.1283530443906784, 0.13818155229091644, -0.00868943240493536, -0.05460423603653908, -0.1013643890619278, -0.10236792266368866, 0.08925779908895493, -0.05773641914129257, 0.07476430386304855, -0.14885357022285461, -0.06675903499126434, 0.015772046521306038, 0.016141414642333984, -0.09562095999717712, 0.02571965754032135, -0.01625603251159191, -0.18119946122169495, 0.056570518761873245, -0.048285093158483505, 0.0440407395362854, -0.06347788125276566, -0.1110161691904068, -0.17226378619670868, 0.06091433763504028, 0.08593481779098511, -0.053876690566539764, -0.12229149043560028, 0.011023230850696564, -0.00012518465518951416, -0.06341652572154999, -0.05023367330431938, 0.09722746908664703, -0.11020902544260025, 0.031452205032110214, -0.012567701749503613, 0.08853451162576675, -0.03510405123233795, -0.011538895778357983, 0.044220831245183945, -0.08039166033267975, -0.009481523185968399, 0.03534642979502678, -0.026372017338871956, -0.04127239063382149, -0.2689029574394226, 0.0036654395516961813, 0.0341104120016098, 0.02497158572077751, 0.07856601476669312, 0.011906822212040424, 0.021174922585487366, 0.03993808850646019, -0.15396519005298615, -0.013395369984209538, 0.14574195444583893, -0.07689505815505981, -0.022186370566487312, 0.05703273415565491, -0.09054436534643173, 0.013882770203053951, -0.030287226662039757, 0.1345842480659485, 0.023923413828015327, 0.06404478847980499, -0.0851147472858429, 0.10106813907623291, -0.1451139897108078, -0.04998219385743141, -0.01244612317532301, 0.09761348366737366, 0.07019034773111343, -0.10272270441055298, 0.014697125181555748, 0.04210108891129494, 0.19416837394237518, 0.016384804621338844, -0.0356343574821949, -0.03396720811724663, 0.004015897400677204, 0.22076453268527985, 0.03044266067445278, 0.10457023978233337, 0.07281364500522614, -0.026583973318338394, 0.12624378502368927, 0.09929762035608292, 0.11280370503664017, -0.055645186454057693, 0.13904185593128204, 0.04667386785149574, 0.038641396909952164, 0.0614289753139019, 0.06836545467376709, 0.09098632633686066, -0.0008288522367365658, 0.1138714924454689, 0.013811973854899406, -0.02422109805047512, -0.021335409954190254, 0.17759373784065247, 0.10501719266176224, -0.14769648015499115, 0.029047364369034767, -0.01258957851678133, 0.039933037012815475, -0.014194529503583908, -0.15634691715240479, -0.07240267097949982, -0.3315149247646332, 0.1226184144616127, -0.07119352370500565, 0.019930170848965645, 0.007913772016763687, -0.037425633519887924, -0.03296699747443199, -0.04477746784687042, 0.13151589035987854, -0.013641550205647945, -0.006079165264964104, -0.04815853759646416, -0.015360191464424133, -0.11607866734266281, -0.11200575530529022, -0.013207737356424332, -0.13671602308750153, -0.010119039565324783, 0.05595948174595833, 0.003977729007601738, 0.01821410097181797, -0.03142618387937546, 0.0024383175186812878, 0.06541839241981506, -0.05751744285225868, 0.056182678788900375, 0.12097269296646118, 0.08766137808561325, -0.1058853268623352, 0.031048951670527458, 0.2011747509241104, 0.04359564557671547, -0.12483977526426315, 0.01449228823184967, 0.1819491684436798, 0.004885740112513304, 0.017068125307559967, -0.006097703706473112, -0.0540788508951664, -0.07554277032613754, 0.1251034289598465, 0.08296554535627365, -0.09985227137804031, 0.015833314508199692, -0.0726347416639328, -0.01594804972410202, -0.06374675035476685, 0.10130585730075836, 0.09538925439119339, 0.04440245032310486, -0.10621760785579681, -0.08487539738416672, -0.10891728103160858, 0.040588874369859695, -0.08629853278398514, -0.07311757653951645, 0.09629398584365845, -0.07057105004787445, -0.07029950618743896, 0.025521177798509598, -0.17978744208812714, -0.009467960335314274, 0.1711762249469757, -0.24654000997543335, -0.0916430801153183, -0.10857923328876495, 0.14477859437465668, 0.016497576609253883, 0.1013975441455841, -0.006207061931490898, -0.007889035157859325, -0.20577777922153473, 0.024890204891562462, -0.05293011665344238, -0.02073732763528824, 0.07814782857894897, -0.09476397186517715, 0.22629831731319427, -0.08276885002851486, 0.020940175279974937, 0.012659613974392414, 0.0870661810040474, -0.030675338581204414, 0.09283176809549332, -0.03660329803824425, -0.12576518952846527, -0.03620953485369682, 0.03001813031733036, 0.013904244638979435, 0.10071761906147003, 0.09772487729787827, -0.03414725139737129, 0.03389119729399681, 0.09747414290904999, 0.04172342270612717, -0.023843804374337196, 0.0360250361263752, -0.17077107727527618, 0.02182629331946373, -0.018498148769140244, -0.06935930997133255, 0.03687669709324837, -0.06603235751390457, 0.1639697551727295, 0.04022442549467087, 0.0670473501086235, -0.036152735352516174, 0.0073931049555540085, -0.014454689808189869, -0.013775371946394444, -0.026180334389209747, -0.17259705066680908, -0.10422050207853317, -0.1347656100988388, -0.012701659463346004, -0.034971047192811966, 0.04591470584273338, 0.023234914988279343, -0.0003200018545612693, -0.014577031135559082, -0.12090865522623062, 0.04360328987240791, 0.11146783083677292, -0.04631396010518074, -0.026193076744675636 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
automatic-speech-recognition
Imran1/ckb
[ "transformers", "safetensors", "whisper", "automatic-speech-recognition", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
2024-02-10T14:07:04+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #whisper #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #whisper #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 45, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #whisper #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.05983767658472061, 0.15663617849349976, -0.00414510490372777, 0.012621625326573849, 0.10675175487995148, 0.00396517850458622, 0.07058298587799072, 0.10818448662757874, -0.014333043247461319, 0.1301925629377365, 0.031459614634513855, 0.10620059072971344, 0.11486424505710602, 0.17755427956581116, -0.00021593451674561948, -0.21627318859100342, 0.06542544066905975, -0.11467250436544418, 0.023902224376797676, 0.1205042228102684, 0.14280648529529572, -0.10782013833522797, 0.0710505023598671, -0.02651231922209263, -0.014152529649436474, -0.030523719266057014, -0.05870387330651283, -0.06662651896476746, 0.06516408175230026, 0.0716853216290474, 0.05976768955588341, 0.02008269540965557, 0.07725182175636292, -0.2948664724826813, 0.018899710848927498, 0.0727730244398117, 0.011833904311060905, 0.06048334762454033, 0.07948420196771622, -0.06289119273424149, 0.12036014348268509, -0.044804252684116364, 0.1532549113035202, 0.07767832279205322, -0.09226784855127335, -0.19217613339424133, -0.0771055743098259, 0.06758320331573486, 0.1468338817358017, 0.056199874728918076, -0.03856382891535759, 0.15159031748771667, -0.09224481880664825, 0.0102085517719388, 0.06493527442216873, -0.07805083692073822, -0.04958232864737511, 0.027303149923682213, 0.08463363349437714, 0.08637925982475281, -0.1273571401834488, -0.012682586908340454, 0.03438213840126991, 0.02163512259721756, 0.09837246686220169, 0.025364719331264496, 0.11626957356929779, 0.027283066883683205, -0.13964000344276428, -0.055175989866256714, 0.12345059961080551, 0.033505070954561234, -0.05288216099143028, -0.23939087986946106, -0.010561608709394932, -0.009556320495903492, -0.03001241944730282, -0.04216838628053665, 0.03810601681470871, -0.029798293486237526, 0.07650589942932129, 0.01746492274105549, -0.07078345119953156, -0.04342244938015938, 0.06982958316802979, 0.07824850082397461, 0.022348513826727867, -0.02065650187432766, 0.028734240680933, 0.10911912471055984, 0.08262593299150467, -0.12154309451580048, -0.06694398820400238, -0.06854734569787979, -0.09466245025396347, -0.0454239584505558, 0.03469004109501839, 0.06703099608421326, 0.057105712592601776, 0.19864854216575623, 0.011600262485444546, 0.05358051881194115, 0.022981496527791023, 0.01298176683485508, 0.07163717597723007, 0.07945776730775833, -0.051690056920051575, -0.1315721571445465, -0.04847193509340286, 0.11824512481689453, 0.008524151518940926, -0.033710937947034836, -0.02968421019613743, 0.0653507187962532, 0.05568600073456764, 0.11161840707063675, 0.07554161548614502, 0.01568971388041973, -0.07114148139953613, -0.043046265840530396, 0.19346864521503448, -0.15610936284065247, 0.021089470013976097, 0.019353056326508522, -0.05417651683092117, -0.022803083062171936, 0.007743596564978361, 0.017318524420261383, -0.02697303518652916, 0.1045108512043953, -0.07085666805505753, -0.032245416194200516, -0.1046156957745552, -0.055557940155267715, 0.03224421665072441, 0.009115081280469894, -0.030819423496723175, -0.042374368757009506, -0.09924564510583878, -0.0756484866142273, 0.06214139610528946, -0.07012778520584106, -0.06952599436044693, -0.028100011870265007, -0.04856603220105171, 0.012879165820777416, 0.0010717154946178198, 0.12350035458803177, -0.03162076696753502, 0.043779097497463226, -0.04884343594312668, 0.06864890456199646, 0.13179735839366913, 0.032575443387031555, -0.07970008254051208, 0.058469612151384354, -0.22937731444835663, 0.11186469346284866, -0.09973006695508957, 0.03430512547492981, -0.15810096263885498, -0.02635045349597931, 0.024752190336585045, 0.033622484654188156, -0.017231743782758713, 0.13669319450855255, -0.2039388120174408, -0.036121536046266556, 0.1721590757369995, -0.1349588930606842, -0.08518610149621964, 0.06643460690975189, -0.055845119059085846, 0.11782421916723251, 0.049206800758838654, -0.014434589073061943, 0.04594586789608002, -0.13173595070838928, -0.025916490703821182, -0.053098164498806, -0.007177549879997969, 0.15609249472618103, 0.06614800542593002, -0.06571528315544128, 0.03145577386021614, 0.02247771993279457, -0.018577884882688522, -0.045781973749399185, -0.03384651243686676, -0.09418359398841858, 0.007437155116349459, -0.07286001741886139, 0.00992972869426012, -0.017532840371131897, -0.08721724897623062, -0.039823103696107864, -0.16453123092651367, -0.00716154370456934, 0.09300678223371506, 0.010935397818684578, -0.02714768424630165, -0.09726624190807343, 0.006592306774109602, 0.01717078872025013, -0.01454078033566475, -0.15828220546245575, -0.0459267795085907, 0.03719138726592064, -0.1820053607225418, 0.03403490409255028, -0.05244239792227745, 0.035954125225543976, 0.03684226796030998, -0.03816571831703186, -0.013848266564309597, 0.020031210035085678, 0.018333489075303078, -0.017020072788000107, -0.2371053695678711, -0.014824622310698032, -0.04800339788198471, 0.16693253815174103, -0.23147691786289215, 0.03312116861343384, 0.07037223875522614, 0.12888941168785095, 0.003875810420140624, -0.0490296445786953, 0.030063113197684288, -0.05199332535266876, -0.044617995619773865, -0.05644122138619423, -0.006168664898723364, -0.030205117538571358, -0.04949198290705681, 0.050275903195142746, -0.19857677817344666, -0.041567981243133545, 0.11094366759061813, 0.06673718988895416, -0.1588216871023178, -0.0695650652050972, -0.03473977744579315, -0.06271405518054962, -0.09103205800056458, -0.05391426756978035, 0.10852089524269104, 0.04763965308666229, 0.048611950129270554, -0.07248158007860184, -0.04900932312011719, 0.007940629497170448, -0.00704985111951828, -0.03555170074105263, 0.08515505492687225, 0.08571629226207733, -0.11543579399585724, 0.09118600934743881, 0.06718818843364716, 0.06912244111299515, 0.0983632430434227, -0.0017782750073820353, -0.09694159775972366, -0.014548503793776035, 0.018360106274485588, 0.01051856018602848, 0.12805555760860443, -0.07398705929517746, 0.03667636960744858, 0.05262641981244087, -0.035613641142845154, 0.01095122192054987, -0.101106658577919, 0.029197964817285538, 0.0282101072371006, -0.003792217466980219, 0.028733761981129646, -0.04522410035133362, 0.020432880148291588, 0.1023864597082138, 0.03395526856184006, 0.027725959196686745, 0.010809014551341534, -0.04075441509485245, -0.11779133975505829, 0.1720944494009018, -0.09817105531692505, -0.25773105025291443, -0.12466797232627869, -0.001978461164981127, 0.045932475477457047, -0.018764600157737732, 0.01608397625386715, -0.053159136325120926, -0.11253257840871811, -0.10541603714227676, 0.019763922318816185, 0.058765511959791183, -0.08840499073266983, -0.052470505237579346, 0.04951007664203644, 0.036848895251750946, -0.12439411878585815, 0.021039357408881187, 0.04023430123925209, -0.059992119669914246, 0.0014880987582728267, 0.07059671729803085, 0.08472984284162521, 0.18226684629917145, 0.022740190848708153, -0.01784367859363556, 0.017296429723501205, 0.23125670850276947, -0.1456713229417801, 0.09739834815263748, 0.1370985060930252, -0.06344101577997208, 0.08623462915420532, 0.21197044849395752, 0.036558255553245544, -0.08882707357406616, 0.037767693400382996, 0.03336544707417488, -0.036437466740608215, -0.2318716198205948, -0.08410470932722092, 0.001480261329561472, -0.08248372375965118, 0.0952354297041893, 0.09051923453807831, 0.11156398802995682, 0.04929385334253311, -0.10106591880321503, -0.07701091468334198, 0.04251527413725853, 0.11516540497541428, -0.006902680266648531, 0.004321529995650053, 0.09879171848297119, -0.029613742604851723, 0.010339556261897087, 0.09523830562829971, 0.0004232692008372396, 0.18618540465831757, 0.04265686497092247, 0.12916190922260284, 0.08458086103200912, 0.05236417427659035, 0.02661769837141037, 0.01322705764323473, 0.031609587371349335, 0.02576516941189766, -0.02334577962756157, -0.09271565079689026, -0.012906024232506752, 0.1415313482284546, 0.04929639771580696, 0.030407944694161415, 0.020662572234869003, -0.03531459718942642, 0.07301895320415497, 0.16116659343242645, 0.011933310888707638, -0.21851851046085358, -0.05515235662460327, 0.07743874937295914, -0.08626089245080948, -0.11299191415309906, -0.0025294655933976173, 0.021754881367087364, -0.17833879590034485, 0.05397404730319977, -0.016486117616295815, 0.10160378366708755, -0.11242987960577011, -0.02206907607614994, 0.04055493697524071, 0.07460751384496689, -0.03305850550532341, 0.07621917128562927, -0.20276865363121033, 0.1373196691274643, 0.008098544552922249, 0.06249339506030083, -0.11230216175317764, 0.08414414525032043, 0.019059745594859123, -0.0036223498173058033, 0.1621086448431015, -0.009664713405072689, -0.09406581521034241, -0.060111574828624725, -0.07602227479219437, -0.012445085681974888, 0.09843466430902481, -0.0939253643155098, 0.08608877658843994, -0.01022840291261673, -0.03214890882372856, -0.007143673487007618, -0.11786875873804092, -0.1394684612751007, -0.183831125497818, 0.05997816100716591, -0.10696699470281601, 0.03344186022877693, -0.10895431786775589, -0.060553617775440216, -0.03646453842520714, 0.19020794332027435, -0.18181639909744263, -0.08386372029781342, -0.14476649463176727, -0.07653295993804932, 0.1361350119113922, -0.04076695069670677, 0.07850751280784607, -0.00008746175444684923, 0.20719517767429352, 0.001825421117246151, -0.00039511307841166854, 0.08349475264549255, -0.09573810547590256, -0.20032998919487, -0.0880952924489975, 0.13964824378490448, 0.12494690716266632, 0.04542626440525055, -0.006928097922354937, 0.027518225833773613, -0.011671899817883968, -0.11464269459247589, 0.02507087029516697, 0.1405206173658371, 0.06840235739946365, 0.04314489662647247, -0.016979211941361427, -0.15606153011322021, -0.10666806995868683, -0.05322869494557381, 0.021586019545793533, 0.17797614634037018, -0.07007403671741486, 0.1621050238609314, 0.16129834949970245, -0.05420130863785744, -0.2030099630355835, 0.02282964438199997, 0.04042449966073036, -0.013990761712193489, 0.03615177795290947, -0.19683793187141418, 0.07753707468509674, 0.016794858500361443, -0.060990821570158005, 0.13549083471298218, -0.1619698405265808, -0.1508903205394745, 0.09218499809503555, 0.06408262252807617, -0.2138945758342743, -0.13302136957645416, -0.10209991782903671, -0.05448025092482567, -0.10983701795339584, 0.08582660555839539, 0.01998555287718773, 0.0000906725981622003, 0.04219266399741173, 0.03161109238862991, 0.021054213866591454, -0.0520465187728405, 0.20073460042476654, 0.0012120193568989635, 0.03459459915757179, -0.08232162147760391, -0.08637090027332306, 0.026973288506269455, -0.05251563340425491, 0.0672052875161171, -0.016655180603265762, 0.0002542635484132916, -0.09922616183757782, -0.06439188867807388, -0.06020424887537956, 0.03343502804636955, -0.08179902285337448, -0.09706422686576843, -0.058388181030750275, 0.10227678716182709, 0.08968468755483627, -0.03377925977110863, -0.06091363728046417, -0.10292473435401917, 0.06651771068572998, 0.22872710227966309, 0.1885143369436264, 0.06312023848295212, -0.07107747346162796, 0.0009368667961098254, -0.023646708577871323, 0.050360288470983505, -0.1945972442626953, 0.046965986490249634, 0.042262639850378036, 0.028454279527068138, 0.12927067279815674, -0.024874795228242874, -0.16607771813869476, -0.04733136296272278, 0.06063033267855644, -0.059542834758758545, -0.18076083064079285, -0.000619421829469502, 0.09315520524978638, -0.15953904390335083, -0.06748805940151215, 0.023891208693385124, -0.020897341892123222, -0.027535755187273026, 0.004573860205709934, 0.0820559412240982, 0.02817925252020359, 0.11291294544935226, 0.06535529345273972, 0.10744494199752808, -0.10965088754892349, 0.08151662349700928, 0.09152320772409439, -0.10730767250061035, 0.02777967043220997, 0.07435369491577148, -0.05882004648447037, -0.03269755467772484, 0.0057791233994066715, 0.07514561712741852, 0.02294853888452053, -0.07087770849466324, -0.0009696646011434495, -0.1182747483253479, 0.06833867728710175, 0.13341592252254486, 0.033248964697122574, -0.0019442925695329905, 0.044254120439291, 0.02532937377691269, -0.08849740773439407, 0.11402047425508499, 0.03831348940730095, 0.031180279329419136, -0.04628003388643265, -0.005872894544154406, 0.04073992744088173, -0.011434492655098438, -0.01770744100213051, -0.03857431188225746, -0.061015255749225616, -0.009887747466564178, -0.1567201316356659, 0.02684243768453598, -0.0771009624004364, 0.00816130917519331, 0.022786233574151993, -0.03996667265892029, -0.005420312751084566, 0.006734060123562813, -0.08264576643705368, -0.03730582818388939, -0.0037628922145813704, 0.1070059984922409, -0.15296638011932373, 0.00852613802999258, 0.09225248545408249, -0.12423861026763916, 0.07808402180671692, -0.0011087276507169008, -0.013306759297847748, 0.02074836567044258, -0.1374569684267044, 0.051461800932884216, -0.006391053553670645, 0.011301612481474876, 0.028202330693602562, -0.19194763898849487, 0.0008063786081038415, -0.04062483087182045, -0.05044460669159889, -0.012731820344924927, -0.05135709419846535, -0.11374296247959137, 0.10732509195804596, 0.023315785452723503, -0.08887150883674622, -0.01889934204518795, 0.045546844601631165, 0.10550197213888168, -0.05122669041156769, 0.13676951825618744, -0.01927841641008854, 0.0586048886179924, -0.1769271343946457, -0.014012092724442482, -0.018402719870209694, 0.013554446399211884, -0.017449822276830673, -0.00605781190097332, 0.0551704466342926, -0.012471658177673817, 0.23972837626934052, -0.027916517108678818, 0.03500373288989067, 0.06697984784841537, 0.016924316063523293, -0.018179070204496384, 0.08486920595169067, 0.05455834046006203, 0.026243781670928, 0.01494054775685072, 0.017568159848451614, -0.051871586591005325, -0.021555433049798012, -0.1424977034330368, 0.07956096529960632, 0.16729016602039337, 0.09009124338626862, -0.008234765380620956, 0.06473081558942795, -0.11607895791530609, -0.07983584702014923, 0.10896016657352448, -0.03711748123168945, -0.0032444922253489494, -0.05700715631246567, 0.1502007693052292, 0.1525147259235382, -0.16814833879470825, 0.06879524886608124, -0.06271831691265106, -0.05224054306745529, -0.11435537785291672, -0.16904489696025848, -0.06866718828678131, -0.035694681107997894, -0.002330650808289647, -0.05624498426914215, 0.07767387479543686, 0.10255347937345505, 0.007528870366513729, 0.0038026864640414715, 0.08233556896448135, -0.037537459284067154, -0.006316144950687885, 0.04542352631688118, 0.049430496990680695, 0.015805410221219063, -0.059124622493982315, 0.010986202396452427, 0.004953318741172552, 0.04692067950963974, 0.05509426072239876, 0.034005217254161835, -0.028324270620942116, 0.012686561793088913, -0.018243486061692238, -0.10028578341007233, 0.035927701741456985, -0.033664118498563766, -0.05780354142189026, 0.13973994553089142, 0.0218597874045372, 0.007779987063258886, -0.02196359448134899, 0.22996114194393158, -0.07252145558595657, -0.08971016108989716, -0.1408918797969818, 0.13730354607105255, -0.046912964433431625, 0.05402535945177078, 0.04905577376484871, -0.10465127229690552, 0.0241316556930542, 0.14292258024215698, 0.13702698051929474, -0.027644719928503036, 0.010874779894948006, 0.015687033534049988, 0.00620539765805006, -0.031101418659090996, 0.04872303828597069, 0.04169761762022972, 0.13120494782924652, -0.06359384953975677, 0.0914405807852745, -0.010274309664964676, -0.08765450119972229, -0.0231675673276186, 0.1299583613872528, 0.005232672207057476, 0.02307419292628765, -0.08125553280115128, 0.11583263427019119, -0.0691702738404274, -0.24996554851531982, 0.04865904897451401, -0.05924736708402634, -0.15156961977481842, -0.017320360988378525, 0.02757420763373375, 0.005632835440337658, 0.02303774654865265, 0.06296881288290024, -0.06651590019464493, 0.1557060331106186, 0.035915885120630264, -0.07977382838726044, -0.06385304778814316, 0.08052598685026169, -0.08511006832122803, 0.29178112745285034, 0.010383724234998226, 0.05882499739527702, 0.0948280319571495, -0.028215935453772545, -0.131154403090477, 0.05278646945953369, 0.0955355316400528, -0.07669185847043991, 0.070269875228405, 0.19858962297439575, 0.0003398389380890876, 0.11546503752470016, 0.07913552224636078, -0.09058261662721634, 0.05968843400478363, -0.07367776334285736, -0.09094593673944473, -0.0922231450676918, 0.08578167855739594, -0.06759190559387207, 0.15170368552207947, 0.12874077260494232, -0.043129127472639084, -0.001158626051619649, -0.030579449608922005, 0.051351167261600494, -0.0008969766786321998, 0.12188339978456497, 0.015837527811527252, -0.19386562705039978, 0.031386423856019974, -0.015537483617663383, 0.099497489631176, -0.23898114264011383, -0.07769263535737991, 0.03750690072774887, -0.014490727335214615, -0.048680152744054794, 0.11743341386318207, 0.05373985692858696, 0.045937854796648026, -0.05465031415224075, -0.060885775834321976, 0.006574091035872698, 0.1611197590827942, -0.11137263476848602, 0.004140520468354225 ]
null
null
peft
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.8.2
{"library_name": "peft", "base_model": "mistralai/Mistral-7B-v0.1"}
null
Devean/Lora-add-paramter
[ "peft", "tensorboard", "safetensors", "arxiv:1910.09700", "base_model:mistralai/Mistral-7B-v0.1", "region:us" ]
2024-02-10T14:10:13+00:00
[ "1910.09700" ]
[]
TAGS #peft #tensorboard #safetensors #arxiv-1910.09700 #base_model-mistralai/Mistral-7B-v0.1 #region-us
# Model Card for Model ID ## Model Details ### Model Description - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact ### Framework versions - PEFT 0.8.2
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ "TAGS\n#peft #tensorboard #safetensors #arxiv-1910.09700 #base_model-mistralai/Mistral-7B-v0.1 #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ 43, 6, 3, 54, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4, 11 ]
[ "passage: TAGS\n#peft #tensorboard #safetensors #arxiv-1910.09700 #base_model-mistralai/Mistral-7B-v0.1 #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2" ]
[ -0.11710534989833832, 0.2021108716726303, -0.00305310171097517, 0.028846031054854393, 0.08059801161289215, 0.015004060231149197, 0.05897754058241844, 0.1303057074546814, 0.031187374144792557, 0.12189128249883652, 0.07221290469169617, 0.12179826200008392, 0.11337313801050186, 0.21017156541347504, 0.0032703992910683155, -0.17037683725357056, 0.017329316586256027, -0.07105738669633865, 0.016411568969488144, 0.11954159289598465, 0.1406661421060562, -0.09835869073867798, 0.07564743608236313, -0.020650992169976234, 0.00018400292901787907, -0.029406584799289703, -0.06564560532569885, -0.012953510507941246, 0.051471225917339325, 0.03191021829843521, 0.06046702340245247, -0.01426337193697691, 0.08398421853780746, -0.26632386445999146, 0.01890237256884575, 0.04357733950018883, -0.0010075353784486651, 0.08578205853700638, 0.1036495789885521, -0.04474060609936714, 0.12027343362569809, -0.02462036907672882, 0.1349533647298813, 0.08896621316671371, -0.09840169548988342, -0.23043224215507507, -0.061313875019550323, 0.08701571822166443, 0.1862124800682068, 0.08349105715751648, -0.044058024883270264, 0.12130947411060333, -0.055914897471666336, 0.023782797157764435, 0.06230388209223747, -0.10898911207914352, -0.06794822961091995, 0.06361687928438187, 0.13435101509094238, 0.07417039573192596, -0.12540510296821594, -0.036571841686964035, 0.03478977456688881, 0.04753020405769348, 0.057528305798769, 0.007320915814489126, 0.1592566967010498, 0.027335582301020622, -0.1491367071866989, -0.04822768270969391, 0.12632375955581665, 0.0060308268293738365, -0.03805458918213844, -0.2204342484474182, -0.0011816532351076603, -0.09180286526679993, -0.03715801611542702, -0.053672049194574356, 0.034506961703300476, 0.01616407185792923, 0.1312389373779297, -0.05237104743719101, -0.09406499564647675, -0.012883263640105724, 0.10434366762638092, 0.05935348942875862, 0.019948996603488922, -0.020193343982100487, 0.005021481309086084, 0.11938370019197464, 0.06570546329021454, -0.13218942284584045, -0.059991396963596344, -0.0678723156452179, -0.03287680447101593, -0.02926778979599476, 0.04342860355973244, 0.0060956296510994434, 0.05442970246076584, 0.271614670753479, -0.029289864003658295, 0.06273429095745087, 0.0413677841424942, 0.024559160694479942, 0.030803659930825233, 0.10673020780086517, -0.035388436168432236, -0.1592811942100525, -0.010240764357149601, 0.10038815438747406, 0.004941240884363651, -0.0340115912258625, -0.05792113021016121, 0.0339377336204052, 0.03688969835639, 0.11829105764627457, 0.1143324226140976, -0.024747367948293686, -0.0769081562757492, -0.05688918009400368, 0.19325868785381317, -0.15363550186157227, 0.04937409237027168, 0.029093055054545403, -0.000040334540244657546, -0.05469810217618942, 0.006924729328602552, 0.017425324767827988, -0.03414385765790939, 0.07773751765489578, -0.06866323202848434, -0.03684751316905022, -0.12194308638572693, -0.03358156606554985, 0.03797125071287155, 0.004391038324683905, -0.047569405287504196, -0.040585316717624664, -0.06721588969230652, -0.1104498878121376, 0.10932566970586777, -0.05606110766530037, -0.05643950775265694, -0.029089007526636124, -0.08311222493648529, 0.02223409153521061, 0.03642607107758522, 0.07339008152484894, -0.023291772231459618, 0.044618409126996994, -0.010840969160199165, 0.06620418280363083, 0.07160625606775284, 0.03322174772620201, -0.08013325184583664, 0.06643705815076828, -0.19171784818172455, 0.07117698341608047, -0.07635278254747391, 0.04034438356757164, -0.16242387890815735, -0.0032556881196796894, -0.0037975399754941463, 0.029654761776328087, 0.043456997722387314, 0.15439356863498688, -0.2152840495109558, -0.028713885694742203, 0.16421760618686676, -0.10821054875850677, -0.13156172633171082, 0.04147589951753616, -0.03590322285890579, 0.18396200239658356, 0.029213888570666313, 0.024624647572636604, 0.0853433758020401, -0.1604994237422943, -0.023839274421334267, -0.02316717430949211, 0.006815835367888212, 0.06776745617389679, 0.08052957057952881, -0.09579490125179291, 0.004500081297010183, 0.008273431099951267, -0.058969367295503616, -0.01605364680290222, -0.038531310856342316, -0.10227826982736588, 0.004596833139657974, -0.0846891850233078, 0.011273767799139023, 0.00774207292124629, -0.09547952562570572, -0.0071614524349570274, -0.15367205440998077, -0.050892043858766556, 0.07809000462293625, 0.001127537339925766, -0.013305325992405415, -0.0938672348856926, 0.06214471533894539, -0.041973941028118134, -0.02055184356868267, -0.1462661772966385, -0.016611823812127113, 0.0156685933470726, -0.13653318583965302, 0.002599505241960287, -0.12617531418800354, 0.06977472454309464, 0.005335449706763029, -0.04819479212164879, -0.041788481175899506, 0.00018171982083003968, -0.005624611396342516, -0.0657259076833725, -0.23649314045906067, -0.02547547034919262, -0.05542780086398125, 0.15732842683792114, -0.2288595288991928, 0.04378710687160492, 0.005164192523807287, 0.11555077135562897, 0.008290127851068974, -0.06289611756801605, 0.025144804269075394, -0.05937067046761513, -0.02614787593483925, -0.07104624062776566, -0.0021156759466975927, 0.0013703849399462342, -0.028434865176677704, 0.018271921202540398, -0.12951606512069702, -0.06610709428787231, 0.09623879939317703, 0.056326109915971756, -0.14839673042297363, 0.006987780332565308, -0.038698066025972366, -0.056878603994846344, -0.06650451570749283, -0.07071413844823837, 0.08863753825426102, 0.054677266627550125, 0.0444474071264267, -0.07999161630868912, -0.06880688667297363, 0.003075228538364172, -0.024279512465000153, -0.019771559163928032, 0.1250922828912735, 0.09412439167499542, -0.09899824112653732, 0.09644327312707901, 0.07666806876659393, 0.022097157314419746, 0.07904145121574402, -0.020689666271209717, -0.10777115821838379, -0.02932608686387539, 0.05866078659892082, 0.01747680827975273, 0.1664396971464157, -0.07203654199838638, 0.054345302283763885, 0.043895915150642395, -0.057049598544836044, 0.04737187549471855, -0.09178517758846283, 0.0073945061303675175, 0.0012111050309613347, -0.015219354070723057, 0.04011654853820801, -0.020511126145720482, 0.004089611582458019, 0.08791087567806244, 0.0637432336807251, 0.02571709454059601, 0.015375961549580097, -0.033283114433288574, -0.1409233659505844, 0.1771393120288849, -0.09036426991224289, -0.23698607087135315, -0.15482057631015778, 0.05546484515070915, 0.059221189469099045, -0.011984648182988167, 0.02799101360142231, -0.05063161998987198, -0.09561813622713089, -0.08848751336336136, 0.010167579166591167, 0.030561476945877075, -0.056560859084129333, -0.06596750766038895, 0.037086062133312225, 0.03745488449931145, -0.11966018378734589, 0.026600511744618416, 0.05526750162243843, -0.005697336047887802, -0.007191151846200228, 0.0506105050444603, 0.09435506165027618, 0.20477622747421265, -0.0007294875686056912, 0.0009280938538722694, 0.055840060114860535, 0.27237534523010254, -0.15663588047027588, 0.11672937124967575, 0.13918790221214294, -0.07324792444705963, 0.07803624868392944, 0.18902404606342316, 0.031991250813007355, -0.09258998185396194, 0.017774123698472977, 0.0313614122569561, -0.02439233846962452, -0.2700929343700409, -0.04677325487136841, -0.02298561856150627, -0.07251652330160141, 0.08538267761468887, 0.08981256932020187, 0.09035968035459518, 0.03318682312965393, -0.06228272616863251, -0.09878776967525482, 0.027313081547617912, 0.11442053318023682, -0.019524384289979935, 0.003035907633602619, 0.08192837238311768, -0.04994804784655571, 0.0046378192491829395, 0.08567038923501968, -0.015649260953068733, 0.13052649796009064, 0.06076815351843834, 0.10975784808397293, 0.07959365844726562, 0.0825527012348175, -0.008604012429714203, 0.03265907242894173, -0.0015131422551348805, 0.020241474732756615, 0.020599620416760445, -0.09041934460401535, 0.020489737391471863, 0.11407609283924103, 0.018341077491641045, 0.022702189162373543, 0.018593423068523407, -0.06288260966539383, 0.03858230635523796, 0.18862862884998322, 0.026827627792954445, -0.20461498200893402, -0.0847124382853508, 0.05392096936702728, -0.07182491570711136, -0.15355747938156128, -0.010881551541388035, 0.019300900399684906, -0.1514490395784378, 0.016321662813425064, -0.04647396504878998, 0.10937470942735672, -0.06718313694000244, -0.041667111217975616, 0.106267049908638, 0.05002044886350632, -0.028089765459299088, 0.04718586802482605, -0.18992765247821808, 0.10674886405467987, 0.030813170596957207, 0.07218228280544281, -0.08872541040182114, 0.0908493772149086, -0.003805613610893488, -0.015122517943382263, 0.1637120246887207, -0.00594486016780138, -0.06529754400253296, -0.08102834224700928, -0.08437960594892502, -0.010874117724597454, 0.08472900092601776, -0.1328064501285553, 0.07531937956809998, -0.033838916569948196, -0.028905747458338737, -0.0042162747122347355, -0.09059762209653854, -0.11798017472028732, -0.16619493067264557, 0.059517573565244675, -0.08593396842479706, 0.016079431399703026, -0.07745571434497833, -0.05434619262814522, 0.03286023065447807, 0.17894059419631958, -0.21505607664585114, -0.10812044143676758, -0.14414986968040466, -0.10078372061252594, 0.1541791409254074, -0.04675482586026192, 0.08446301519870758, -0.007559371646493673, 0.16159680485725403, -0.0034639653749763966, -0.018588881939649582, 0.08569793403148651, -0.0931713655591011, -0.18468298017978668, -0.04723035916686058, 0.17987769842147827, 0.13188622891902924, 0.025678735226392746, -0.011944744735956192, 0.031038351356983185, -0.06527556478977203, -0.10952436923980713, 0.03487575799226761, 0.1461389809846878, 0.06625843048095703, -0.026274340227246284, -0.04168965667486191, -0.1043180599808693, -0.06579500436782837, -0.0404425784945488, -0.004437439609318972, 0.2012956440448761, -0.07110010087490082, 0.1533629298210144, 0.1086113229393959, -0.06448375433683395, -0.21142131090164185, 0.03830984979867935, 0.039698924869298935, 0.015612287446856499, 0.03754628822207451, -0.19021426141262054, 0.08549356460571289, -0.023149000480771065, -0.08021710067987442, 0.18305234611034393, -0.2002161145210266, -0.13059002161026, 0.1032833531498909, 0.024186357855796814, -0.2063939869403839, -0.15295074880123138, -0.10558506846427917, -0.011016320437192917, -0.10945338755846024, 0.04749727249145508, 0.015990063548088074, 0.011770597659051418, 0.013257361017167568, 0.02004881761968136, 0.04308827966451645, -0.0494568757712841, 0.2023766189813614, -0.047972604632377625, -0.006063458509743214, -0.05612028017640114, -0.08288472145795822, 0.01696155034005642, -0.053984127938747406, 0.12219367176294327, -0.018605204299092293, 0.02843589335680008, -0.157892107963562, -0.044674523174762726, -0.05601145699620247, 0.03513455390930176, -0.09516076743602753, -0.07899542897939682, -0.042643509805202484, 0.08191783726215363, 0.08799072355031967, -0.014260710217058659, 0.016456549987196922, -0.09486821293830872, 0.09404904395341873, 0.20166359841823578, 0.18991129100322723, 0.06911168992519379, -0.059349436312913895, 0.03237222135066986, -0.039019353687763214, 0.044231124222278595, -0.21585260331630707, 0.04015108197927475, 0.06397762894630432, 0.0214134082198143, 0.07001511752605438, -0.0033791514579206705, -0.1601126790046692, -0.08526817709207535, 0.08770349621772766, -0.05913963168859482, -0.17669260501861572, -0.029871204867959023, 0.04047708585858345, -0.21304070949554443, -0.047274958342313766, 0.037410225719213486, -0.01852087862789631, -0.04006629064679146, 0.027995970100164413, 0.08235564827919006, -0.019419431686401367, 0.08557318896055222, 0.0928596630692482, 0.0853390321135521, -0.08894116431474686, 0.05364793911576271, 0.08021722733974457, -0.030690550804138184, 0.031201256439089775, 0.1374773532152176, -0.0313359759747982, -0.046540141105651855, 0.08343418687582016, 0.11911316215991974, 0.0004875446029473096, -0.050768740475177765, 0.006001137662678957, -0.04436188191175461, 0.0638790875673294, 0.12114119529724121, 0.018999839201569557, -0.009910540655255318, 0.07356899976730347, 0.02887750044465065, -0.09400276839733124, 0.12378161400556564, 0.04546736925840378, 0.023448480293154716, -0.037249933928251266, -0.02375900372862816, -0.01419738493859768, -0.008447879925370216, -0.015033082105219364, 0.0019479027250781655, -0.09141910076141357, -0.0018762131221592426, -0.10813231021165848, 0.020840857177972794, -0.07142873108386993, -0.0016835704445838928, 0.028542915359139442, -0.04929929971694946, -0.0021413906943053007, -0.00535619119182229, -0.07493800669908524, -0.05454593151807785, -0.0220900047570467, 0.07482270896434784, -0.1443692445755005, 0.032480642199516296, 0.07352779805660248, -0.10033171623945236, 0.06492601335048676, -0.010404354892671108, 0.012511092238128185, 0.001769888331182301, -0.1525413691997528, 0.06084395572543144, -0.027755670249462128, -0.009639757685363293, -0.0018469417700544, -0.1788497418165207, -0.00968952290713787, -0.04198571667075157, -0.0719187781214714, 0.014900799840688705, -0.005501140840351582, -0.12420886009931564, 0.1125902310013771, 0.009114287793636322, -0.06442797183990479, -0.014777983538806438, 0.04569220915436745, 0.0749797374010086, -0.012786103412508965, 0.11202658712863922, -0.03053564578294754, 0.08561937510967255, -0.18019911646842957, -0.0029284334741532803, -0.01648585870862007, 0.05532195791602135, -0.018937058746814728, -0.043604105710983276, 0.054213520139455795, -0.02149723842740059, 0.15956193208694458, 0.0006230112048797309, 0.06613242626190186, 0.05339767038822174, 0.012420245446264744, 0.044190000742673874, 0.07510527968406677, 0.06446485221385956, -0.01652139611542225, -0.008392492309212685, 0.028617747128009796, 0.001371529302559793, -0.04747578501701355, -0.1385776698589325, 0.06883858144283295, 0.17621056735515594, 0.06481525301933289, 0.024277377873659134, 0.009303689934313297, -0.12749531865119934, -0.07630964368581772, 0.10874603688716888, -0.016498900949954987, -0.03471783548593521, -0.06592123210430145, 0.22797319293022156, 0.1479867696762085, -0.18833231925964355, 0.07474634051322937, -0.050895243883132935, -0.03557534143328667, -0.14153027534484863, -0.1760072559118271, -0.05657065659761429, -0.041112907230854034, -0.03338297829031944, -0.058557938784360886, 0.05158796161413193, 0.03337971493601799, -0.005050627049058676, -0.018989838659763336, 0.10635516792535782, 0.02567315846681595, -0.03679250180721283, 0.04605171084403992, 0.06206009164452553, 0.042398180812597275, -0.09936370700597763, 0.014429115690290928, 0.001129169948399067, 0.008629530668258667, 0.06442737579345703, 0.025388963520526886, -0.06442403793334961, 0.02965252660214901, -0.021092981100082397, -0.12110313028097153, 0.04773139953613281, -0.005093853455036879, -0.02447228878736496, 0.15094423294067383, 0.035077501088380814, 0.0069218906573951244, -0.014097390696406364, 0.2359100878238678, -0.06959674507379532, -0.07762467861175537, -0.12911878526210785, 0.07983440160751343, -0.06683376431465149, 0.02375158481299877, 0.010837567038834095, -0.12336385995149612, 0.017015453428030014, 0.1758856624364853, 0.12080017477273941, -0.013587837107479572, 0.011423238553106785, 0.044204335659742355, 0.010076981037855148, -0.031474411487579346, 0.01151349302381277, 0.05340633913874626, 0.20649796724319458, -0.07732890546321869, 0.058807097375392914, -0.01771836169064045, -0.06702535599470139, -0.03160260617733002, 0.10637380927801132, -0.012023967690765858, -0.008326183073222637, -0.06214261054992676, 0.14579035341739655, -0.0741109848022461, -0.20746080577373505, 0.05533681809902191, -0.08247096091508865, -0.13790784776210785, -0.046441394835710526, 0.036611612886190414, -0.0293093491345644, 0.008891514502465725, 0.06426571309566498, -0.049631714820861816, 0.17413274943828583, 0.030408212915062904, -0.048054393380880356, -0.09097152203321457, 0.05287618935108185, -0.15801310539245605, 0.28309598565101624, 0.02234046161174774, 0.04158395528793335, 0.11151735484600067, -0.020557766780257225, -0.13807201385498047, 0.010151871480047703, 0.10906568914651871, -0.05942131578922272, 0.06637952476739883, 0.15963231027126312, 0.0020506156142801046, 0.12095240503549576, 0.06462416797876358, -0.05576562508940697, 0.03656291216611862, -0.07868297398090363, -0.055026356130838394, -0.11541097611188889, 0.0790775939822197, -0.09621909260749817, 0.15381580591201782, 0.1203673854470253, -0.07174723595380783, -0.005297505762428045, -0.02115396410226822, 0.07766447961330414, 0.018183840438723564, 0.11580948531627655, 0.007772176526486874, -0.18437495827674866, 0.04190204665064812, 0.0063735381700098515, 0.10094809532165527, -0.2102777361869812, -0.052648842334747314, 0.04323260858654976, -0.01908690482378006, -0.08854816854000092, 0.1109946221113205, 0.042083997279405594, 0.01879647560417652, -0.0373709499835968, -0.05093289539217949, 0.021104082465171814, 0.1511654555797577, -0.10028860718011856, -0.01127503253519535 ]
null
null
transformers
# Model Card ## Summary This model was trained using [H2O LLM Studio](https://github.com/h2oai/h2o-llmstudio). - Base model: [deepseek-ai/deepseek-coder-1.3b-instruct](https://huggingface.co/deepseek-ai/deepseek-coder-1.3b-instruct) ## Usage To use the model with the `transformers` library on a machine with GPUs, first make sure you have the `transformers` library installed. ```bash pip install transformers==4.36.1 ``` Also make sure you are providing your huggingface token to the pipeline if the model is lying in a private repo. - Either leave `token=True` in the `pipeline` and login to hugginface_hub by running ```python import huggingface_hub huggingface_hub.login(<ACCESS_TOKEN>) ``` - Or directly pass your <ACCESS_TOKEN> to `token` in the `pipeline` ```python from transformers import pipeline generate_text = pipeline( model="mwalol/json-deepseek-v2-1-1", torch_dtype="auto", trust_remote_code=True, use_fast=True, device_map={"": "cuda:0"}, token=True, ) res = generate_text( "Why is drinking water so healthy?", min_new_tokens=0, max_new_tokens=1, do_sample=False, num_beams=1, temperature=float(0.0), repetition_penalty=float(1.2), renormalize_logits=True ) print(res[0]["generated_text"]) ``` You can print a sample prompt after the preprocessing step to see how it is feed to the tokenizer: ```python print(generate_text.preprocess("Why is drinking water so healthy?")["prompt_text"]) ``` ```bash <|user|>Why is drinking water so healthy?<|EOT|><|assistant|> ``` Alternatively, you can download [h2oai_pipeline.py](h2oai_pipeline.py), store it alongside your notebook, and construct the pipeline yourself from the loaded model and tokenizer. If the model and the tokenizer are fully supported in the `transformers` package, this will allow you to set `trust_remote_code=False`. ```python from h2oai_pipeline import H2OTextGenerationPipeline from transformers import AutoModelForCausalLM, AutoTokenizer tokenizer = AutoTokenizer.from_pretrained( "mwalol/json-deepseek-v2-1-1", use_fast=True, padding_side="left", trust_remote_code=True, ) model = AutoModelForCausalLM.from_pretrained( "mwalol/json-deepseek-v2-1-1", torch_dtype="auto", device_map={"": "cuda:0"}, trust_remote_code=True, ) generate_text = H2OTextGenerationPipeline(model=model, tokenizer=tokenizer) res = generate_text( "Why is drinking water so healthy?", min_new_tokens=0, max_new_tokens=1, do_sample=False, num_beams=1, temperature=float(0.0), repetition_penalty=float(1.2), renormalize_logits=True ) print(res[0]["generated_text"]) ``` You may also construct the pipeline from the loaded model and tokenizer yourself and consider the preprocessing steps: ```python from transformers import AutoModelForCausalLM, AutoTokenizer model_name = "mwalol/json-deepseek-v2-1-1" # either local folder or huggingface model name # Important: The prompt needs to be in the same format the model was trained with. # You can find an example prompt in the experiment logs. prompt = "<|user|>How are you?<|EOT|><|assistant|>" tokenizer = AutoTokenizer.from_pretrained( model_name, use_fast=True, trust_remote_code=True, ) model = AutoModelForCausalLM.from_pretrained( model_name, torch_dtype="auto", device_map={"": "cuda:0"}, trust_remote_code=True, ) model.cuda().eval() inputs = tokenizer(prompt, return_tensors="pt", add_special_tokens=False).to("cuda") # generate configuration can be modified to your needs tokens = model.generate( input_ids=inputs["input_ids"], attention_mask=inputs["attention_mask"], min_new_tokens=0, max_new_tokens=1, do_sample=False, num_beams=1, temperature=float(0.0), repetition_penalty=float(1.2), renormalize_logits=True )[0] tokens = tokens[inputs["input_ids"].shape[1]:] answer = tokenizer.decode(tokens, skip_special_tokens=True) print(answer) ``` ## Quantization and sharding You can load the models using quantization by specifying ```load_in_8bit=True``` or ```load_in_4bit=True```. Also, sharding on multiple GPUs is possible by setting ```device_map=auto```. ## Model Architecture ``` LlamaForCausalLM( (model): LlamaModel( (embed_tokens): Embedding(32256, 2048, padding_idx=32014) (layers): ModuleList( (0-23): 24 x LlamaDecoderLayer( (self_attn): LlamaFlashAttention2( (q_proj): Linear(in_features=2048, out_features=2048, bias=False) (k_proj): Linear(in_features=2048, out_features=2048, bias=False) (v_proj): Linear(in_features=2048, out_features=2048, bias=False) (o_proj): Linear(in_features=2048, out_features=2048, bias=False) (rotary_emb): LlamaLinearScalingRotaryEmbedding() ) (mlp): LlamaMLP( (gate_proj): Linear(in_features=2048, out_features=5504, bias=False) (up_proj): Linear(in_features=2048, out_features=5504, bias=False) (down_proj): Linear(in_features=5504, out_features=2048, bias=False) (act_fn): SiLU() ) (input_layernorm): LlamaRMSNorm() (post_attention_layernorm): LlamaRMSNorm() ) ) (norm): LlamaRMSNorm() ) (lm_head): Linear(in_features=2048, out_features=32256, bias=False) ) ``` ## Model Configuration This model was trained using H2O LLM Studio and with the configuration in [cfg.yaml](cfg.yaml). Visit [H2O LLM Studio](https://github.com/h2oai/h2o-llmstudio) to learn how to train your own large language models. ## Disclaimer Please read this disclaimer carefully before using the large language model provided in this repository. Your use of the model signifies your agreement to the following terms and conditions. - Biases and Offensiveness: The large language model is trained on a diverse range of internet text data, which may contain biased, racist, offensive, or otherwise inappropriate content. By using this model, you acknowledge and accept that the generated content may sometimes exhibit biases or produce content that is offensive or inappropriate. The developers of this repository do not endorse, support, or promote any such content or viewpoints. - Limitations: The large language model is an AI-based tool and not a human. It may produce incorrect, nonsensical, or irrelevant responses. It is the user's responsibility to critically evaluate the generated content and use it at their discretion. - Use at Your Own Risk: Users of this large language model must assume full responsibility for any consequences that may arise from their use of the tool. The developers and contributors of this repository shall not be held liable for any damages, losses, or harm resulting from the use or misuse of the provided model. - Ethical Considerations: Users are encouraged to use the large language model responsibly and ethically. By using this model, you agree not to use it for purposes that promote hate speech, discrimination, harassment, or any form of illegal or harmful activities. - Reporting Issues: If you encounter any biased, offensive, or otherwise inappropriate content generated by the large language model, please report it to the repository maintainers through the provided channels. Your feedback will help improve the model and mitigate potential issues. - Changes to this Disclaimer: The developers of this repository reserve the right to modify or update this disclaimer at any time without prior notice. It is the user's responsibility to periodically review the disclaimer to stay informed about any changes. By using the large language model provided in this repository, you agree to accept and comply with the terms and conditions outlined in this disclaimer. If you do not agree with any part of this disclaimer, you should refrain from using the model and any content generated by it.
{"language": ["en"], "library_name": "transformers", "tags": ["gpt", "llm", "large language model", "h2o-llmstudio"], "inference": false, "thumbnail": "https://h2o.ai/etc.clientlibs/h2o/clientlibs/clientlib-site/resources/images/favicon.ico"}
text-generation
mwalol/json-deepseek-v2-1-1
[ "transformers", "safetensors", "llama", "text-generation", "gpt", "llm", "large language model", "h2o-llmstudio", "conversational", "en", "autotrain_compatible", "text-generation-inference", "region:us" ]
2024-02-10T14:16:37+00:00
[]
[ "en" ]
TAGS #transformers #safetensors #llama #text-generation #gpt #llm #large language model #h2o-llmstudio #conversational #en #autotrain_compatible #text-generation-inference #region-us
# Model Card ## Summary This model was trained using H2O LLM Studio. - Base model: deepseek-ai/deepseek-coder-1.3b-instruct ## Usage To use the model with the 'transformers' library on a machine with GPUs, first make sure you have the 'transformers' library installed. Also make sure you are providing your huggingface token to the pipeline if the model is lying in a private repo. - Either leave 'token=True' in the 'pipeline' and login to hugginface_hub by running - Or directly pass your <ACCESS_TOKEN> to 'token' in the 'pipeline' You can print a sample prompt after the preprocessing step to see how it is feed to the tokenizer: Alternatively, you can download h2oai_pipeline.py, store it alongside your notebook, and construct the pipeline yourself from the loaded model and tokenizer. If the model and the tokenizer are fully supported in the 'transformers' package, this will allow you to set 'trust_remote_code=False'. You may also construct the pipeline from the loaded model and tokenizer yourself and consider the preprocessing steps: ## Quantization and sharding You can load the models using quantization by specifying or . Also, sharding on multiple GPUs is possible by setting . ## Model Architecture ## Model Configuration This model was trained using H2O LLM Studio and with the configuration in URL. Visit H2O LLM Studio to learn how to train your own large language models. ## Disclaimer Please read this disclaimer carefully before using the large language model provided in this repository. Your use of the model signifies your agreement to the following terms and conditions. - Biases and Offensiveness: The large language model is trained on a diverse range of internet text data, which may contain biased, racist, offensive, or otherwise inappropriate content. By using this model, you acknowledge and accept that the generated content may sometimes exhibit biases or produce content that is offensive or inappropriate. The developers of this repository do not endorse, support, or promote any such content or viewpoints. - Limitations: The large language model is an AI-based tool and not a human. It may produce incorrect, nonsensical, or irrelevant responses. It is the user's responsibility to critically evaluate the generated content and use it at their discretion. - Use at Your Own Risk: Users of this large language model must assume full responsibility for any consequences that may arise from their use of the tool. The developers and contributors of this repository shall not be held liable for any damages, losses, or harm resulting from the use or misuse of the provided model. - Ethical Considerations: Users are encouraged to use the large language model responsibly and ethically. By using this model, you agree not to use it for purposes that promote hate speech, discrimination, harassment, or any form of illegal or harmful activities. - Reporting Issues: If you encounter any biased, offensive, or otherwise inappropriate content generated by the large language model, please report it to the repository maintainers through the provided channels. Your feedback will help improve the model and mitigate potential issues. - Changes to this Disclaimer: The developers of this repository reserve the right to modify or update this disclaimer at any time without prior notice. It is the user's responsibility to periodically review the disclaimer to stay informed about any changes. By using the large language model provided in this repository, you agree to accept and comply with the terms and conditions outlined in this disclaimer. If you do not agree with any part of this disclaimer, you should refrain from using the model and any content generated by it.
[ "# Model Card", "## Summary\n\nThis model was trained using H2O LLM Studio.\n- Base model: deepseek-ai/deepseek-coder-1.3b-instruct", "## Usage\n\nTo use the model with the 'transformers' library on a machine with GPUs, first make sure you have the 'transformers' library installed.\n\n\n\nAlso make sure you are providing your huggingface token to the pipeline if the model is lying in a private repo.\n - Either leave 'token=True' in the 'pipeline' and login to hugginface_hub by running\n \n - Or directly pass your <ACCESS_TOKEN> to 'token' in the 'pipeline'\n\n\n\nYou can print a sample prompt after the preprocessing step to see how it is feed to the tokenizer:\n\n\n\n\n\nAlternatively, you can download h2oai_pipeline.py, store it alongside your notebook, and construct the pipeline yourself from the loaded model and tokenizer. If the model and the tokenizer are fully supported in the 'transformers' package, this will allow you to set 'trust_remote_code=False'.\n\n\n\n\nYou may also construct the pipeline from the loaded model and tokenizer yourself and consider the preprocessing steps:", "## Quantization and sharding\n\nYou can load the models using quantization by specifying or . Also, sharding on multiple GPUs is possible by setting .", "## Model Architecture", "## Model Configuration\n\nThis model was trained using H2O LLM Studio and with the configuration in URL. Visit H2O LLM Studio to learn how to train your own large language models.", "## Disclaimer\n\nPlease read this disclaimer carefully before using the large language model provided in this repository. Your use of the model signifies your agreement to the following terms and conditions.\n\n- Biases and Offensiveness: The large language model is trained on a diverse range of internet text data, which may contain biased, racist, offensive, or otherwise inappropriate content. By using this model, you acknowledge and accept that the generated content may sometimes exhibit biases or produce content that is offensive or inappropriate. The developers of this repository do not endorse, support, or promote any such content or viewpoints.\n- Limitations: The large language model is an AI-based tool and not a human. It may produce incorrect, nonsensical, or irrelevant responses. It is the user's responsibility to critically evaluate the generated content and use it at their discretion.\n- Use at Your Own Risk: Users of this large language model must assume full responsibility for any consequences that may arise from their use of the tool. The developers and contributors of this repository shall not be held liable for any damages, losses, or harm resulting from the use or misuse of the provided model.\n- Ethical Considerations: Users are encouraged to use the large language model responsibly and ethically. By using this model, you agree not to use it for purposes that promote hate speech, discrimination, harassment, or any form of illegal or harmful activities.\n- Reporting Issues: If you encounter any biased, offensive, or otherwise inappropriate content generated by the large language model, please report it to the repository maintainers through the provided channels. Your feedback will help improve the model and mitigate potential issues.\n- Changes to this Disclaimer: The developers of this repository reserve the right to modify or update this disclaimer at any time without prior notice. It is the user's responsibility to periodically review the disclaimer to stay informed about any changes.\n\nBy using the large language model provided in this repository, you agree to accept and comply with the terms and conditions outlined in this disclaimer. If you do not agree with any part of this disclaimer, you should refrain from using the model and any content generated by it." ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #gpt #llm #large language model #h2o-llmstudio #conversational #en #autotrain_compatible #text-generation-inference #region-us \n", "# Model Card", "## Summary\n\nThis model was trained using H2O LLM Studio.\n- Base model: deepseek-ai/deepseek-coder-1.3b-instruct", "## Usage\n\nTo use the model with the 'transformers' library on a machine with GPUs, first make sure you have the 'transformers' library installed.\n\n\n\nAlso make sure you are providing your huggingface token to the pipeline if the model is lying in a private repo.\n - Either leave 'token=True' in the 'pipeline' and login to hugginface_hub by running\n \n - Or directly pass your <ACCESS_TOKEN> to 'token' in the 'pipeline'\n\n\n\nYou can print a sample prompt after the preprocessing step to see how it is feed to the tokenizer:\n\n\n\n\n\nAlternatively, you can download h2oai_pipeline.py, store it alongside your notebook, and construct the pipeline yourself from the loaded model and tokenizer. If the model and the tokenizer are fully supported in the 'transformers' package, this will allow you to set 'trust_remote_code=False'.\n\n\n\n\nYou may also construct the pipeline from the loaded model and tokenizer yourself and consider the preprocessing steps:", "## Quantization and sharding\n\nYou can load the models using quantization by specifying or . Also, sharding on multiple GPUs is possible by setting .", "## Model Architecture", "## Model Configuration\n\nThis model was trained using H2O LLM Studio and with the configuration in URL. Visit H2O LLM Studio to learn how to train your own large language models.", "## Disclaimer\n\nPlease read this disclaimer carefully before using the large language model provided in this repository. Your use of the model signifies your agreement to the following terms and conditions.\n\n- Biases and Offensiveness: The large language model is trained on a diverse range of internet text data, which may contain biased, racist, offensive, or otherwise inappropriate content. By using this model, you acknowledge and accept that the generated content may sometimes exhibit biases or produce content that is offensive or inappropriate. The developers of this repository do not endorse, support, or promote any such content or viewpoints.\n- Limitations: The large language model is an AI-based tool and not a human. It may produce incorrect, nonsensical, or irrelevant responses. It is the user's responsibility to critically evaluate the generated content and use it at their discretion.\n- Use at Your Own Risk: Users of this large language model must assume full responsibility for any consequences that may arise from their use of the tool. The developers and contributors of this repository shall not be held liable for any damages, losses, or harm resulting from the use or misuse of the provided model.\n- Ethical Considerations: Users are encouraged to use the large language model responsibly and ethically. By using this model, you agree not to use it for purposes that promote hate speech, discrimination, harassment, or any form of illegal or harmful activities.\n- Reporting Issues: If you encounter any biased, offensive, or otherwise inappropriate content generated by the large language model, please report it to the repository maintainers through the provided channels. Your feedback will help improve the model and mitigate potential issues.\n- Changes to this Disclaimer: The developers of this repository reserve the right to modify or update this disclaimer at any time without prior notice. It is the user's responsibility to periodically review the disclaimer to stay informed about any changes.\n\nBy using the large language model provided in this repository, you agree to accept and comply with the terms and conditions outlined in this disclaimer. If you do not agree with any part of this disclaimer, you should refrain from using the model and any content generated by it." ]
[ 64, 3, 38, 244, 34, 4, 42, 518 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #gpt #llm #large language model #h2o-llmstudio #conversational #en #autotrain_compatible #text-generation-inference #region-us \n# Model Card## Summary\n\nThis model was trained using H2O LLM Studio.\n- Base model: deepseek-ai/deepseek-coder-1.3b-instruct## Usage\n\nTo use the model with the 'transformers' library on a machine with GPUs, first make sure you have the 'transformers' library installed.\n\n\n\nAlso make sure you are providing your huggingface token to the pipeline if the model is lying in a private repo.\n - Either leave 'token=True' in the 'pipeline' and login to hugginface_hub by running\n \n - Or directly pass your <ACCESS_TOKEN> to 'token' in the 'pipeline'\n\n\n\nYou can print a sample prompt after the preprocessing step to see how it is feed to the tokenizer:\n\n\n\n\n\nAlternatively, you can download h2oai_pipeline.py, store it alongside your notebook, and construct the pipeline yourself from the loaded model and tokenizer. If the model and the tokenizer are fully supported in the 'transformers' package, this will allow you to set 'trust_remote_code=False'.\n\n\n\n\nYou may also construct the pipeline from the loaded model and tokenizer yourself and consider the preprocessing steps:## Quantization and sharding\n\nYou can load the models using quantization by specifying or . Also, sharding on multiple GPUs is possible by setting .## Model Architecture## Model Configuration\n\nThis model was trained using H2O LLM Studio and with the configuration in URL. Visit H2O LLM Studio to learn how to train your own large language models." ]
[ -0.09606858342885971, 0.19560745358467102, -0.005736880004405975, 0.05329291149973869, 0.05006769672036171, 0.006951822433620691, 0.11371155828237534, 0.1088312417268753, 0.07669293135404587, 0.09452244639396667, -0.002002003835514188, -0.00575899425894022, 0.07589101791381836, 0.13600042462348938, 0.07750868797302246, -0.21565362811088562, 0.00875082891434431, -0.09655000269412994, -0.0313391238451004, 0.06301282346248627, 0.041830550879240036, -0.05266651138663292, 0.09501102566719055, 0.023580310866236687, -0.0020056175999343395, -0.001091262442059815, -0.0036409194581210613, 0.031056109815835953, 0.05325499549508095, 0.08127935230731964, 0.013626850210130215, 0.014144389890134335, 0.05630922690033913, -0.13374216854572296, 0.03868100047111511, 0.11743557453155518, 0.007681848481297493, 0.04794013500213623, 0.03502948209643364, -0.03869011253118515, 0.08466466516256332, -0.050512976944446564, 0.04594339802861214, 0.04016922786831856, -0.08598228543996811, -0.14486266672611237, -0.04946830868721008, 0.04228098690509796, 0.13376814126968384, 0.048748504370450974, -0.00002006825707212556, 0.10484185069799423, 0.07275719940662384, 0.07183379679918289, 0.15852977335453033, -0.12929928302764893, -0.03336906060576439, 0.03698194772005081, 0.04581783339381218, 0.10817548632621765, -0.0442771315574646, -0.012174897827208042, -0.009388185106217861, 0.002738380338996649, 0.03211090341210365, -0.06399228423833847, 0.06185347959399223, -0.04329195246100426, -0.11828534305095673, -0.008289942517876625, 0.11654704809188843, -0.06557802110910416, -0.04406295716762543, -0.10969982296228409, -0.14129100739955902, -0.03992059826850891, 0.0303728599101305, -0.015905005857348442, 0.039921898394823074, -0.01518133282661438, 0.07365814596414566, -0.14692629873752594, -0.0934566855430603, -0.08078690618276596, 0.004249582067131996, 0.06813991814851761, 0.04958444461226463, 0.027346793562173843, -0.07809865474700928, 0.2021075338125229, -0.011351630091667175, -0.11853999644517899, -0.07115595042705536, -0.04371286556124687, -0.13050317764282227, 0.0018373077036812901, -0.002111744601279497, -0.07166700810194016, 0.04119908809661865, 0.18604421615600586, -0.01585341803729534, 0.10219910740852356, -0.008821605704724789, 0.009978710673749447, 0.06444813311100006, 0.12773631513118744, -0.04684564843773842, -0.008084750734269619, 0.050941962748765945, 0.00485632847994566, 0.051854442805051804, -0.061545480042696, -0.04352562874555588, -0.033070337027311325, -0.0620734728872776, 0.08267182111740112, 0.025951171293854713, 0.04002057760953903, 0.015801317989826202, -0.0664655938744545, 0.1121671125292778, -0.17230741679668427, 0.05356789380311966, 0.03905847296118736, -0.022832369431853294, -0.006983745377510786, 0.11947596818208694, -0.045249782502651215, -0.09794136881828308, -0.002322576707229018, -0.02122468315064907, 0.022736670449376106, -0.10323237627744675, -0.0410727858543396, -0.01052775140851736, 0.0292768944054842, -0.045772209763526917, -0.0855882540345192, -0.13485966622829437, -0.020572451874613762, 0.09023799002170563, -0.04085138440132141, 0.04009721428155899, 0.002811640268191695, 0.07162667065858841, -0.024276575073599815, 0.03057747147977352, -0.06473319977521896, -0.02255559153854847, -0.014187018387019634, 0.012443482875823975, 0.058588653802871704, -0.010146886110305786, -0.00969663169234991, -0.06387045234441757, 0.10078117251396179, -0.18584044277668, 0.08615193516016006, -0.02765411138534546, 0.005118547007441521, -0.06156293302774429, -0.009667927399277687, -0.020518938079476357, -0.011880946345627308, 0.00880772527307272, 0.06236302852630615, -0.15670417249202728, 0.003959662280976772, 0.18038694560527802, -0.14680339395999908, -0.026688549667596817, 0.07731787115335464, -0.0180928036570549, 0.12444345653057098, 0.05240675061941147, 0.0916505753993988, 0.2780934274196625, -0.2505964934825897, 0.0516829714179039, 0.07189343124628067, -0.09899799525737762, 0.08530008047819138, 0.03666726127266884, 0.008894302882254124, 0.03315553069114685, 0.03142526000738144, -0.08051086962223053, -0.03643609583377838, 0.07074024528265, -0.04137293994426727, -0.009666608646512032, -0.006340096704661846, -0.0567532442510128, -0.037875380367040634, -0.043006349354982376, 0.033667005598545074, -0.06307943165302277, 0.0641600489616394, 0.13971051573753357, -0.08362124860286713, 0.04264967888593674, -0.07578355073928833, 0.07432147115468979, 0.024961235001683235, -0.015422975644469261, -0.10640281438827515, -0.06533747166395187, 0.037627849727869034, -0.20363418757915497, 0.07400663942098618, 0.042633090168237686, 0.02226453274488449, 0.16888616979122162, 0.024444447830319405, -0.017135148867964745, 0.05367368832230568, -0.028682438656687737, -0.0173420999199152, -0.08626700937747955, -0.020074665546417236, -0.03443695232272148, 0.1524539738893509, -0.025225883349776268, 0.08055371791124344, 0.015573062002658844, 0.05463431030511856, 0.022511644288897514, -0.020387491211295128, 0.036379069089889526, -0.09335773438215256, 0.01887381821870804, -0.0642029196023941, 0.008192851208150387, 0.05073123052716255, -0.015898652374744415, 0.01571393571794033, -0.22062428295612335, -0.14346785843372345, 0.04985560104250908, 0.14443163573741913, -0.03435661271214485, -0.11020930111408234, 0.001468116999603808, -0.06180844083428383, -0.051132071763277054, -0.06167202070355415, 0.13914640247821808, 0.048569850623607635, 0.09614482522010803, -0.1128711849451065, -0.07708461582660675, -0.017854295670986176, -0.09431089460849762, 0.025822201743721962, 0.06830307096242905, -0.02701813355088234, -0.0721135064959526, 0.0007475498714484274, 0.0061747306026518345, -0.05535958707332611, 0.17149916291236877, 0.06523136794567108, -0.07091296464204788, -0.07754148542881012, 0.036609284579753876, 0.007417021319270134, 0.08068936318159103, 0.004512880463153124, 0.03852257505059242, 0.024752099066972733, 0.0055897473357617855, 0.0346093475818634, -0.10061394423246384, 0.06292611360549927, 0.01642065867781639, -0.010834166780114174, 0.004465427249670029, 0.07928033918142319, -0.0347716398537159, 0.015739066526293755, -0.030183225870132446, 0.14341013133525848, -0.03498683497309685, -0.09831596910953522, -0.0826975628733635, 0.13153530657291412, -0.10949938744306564, -0.23397715389728546, -0.16176246106624603, -0.08813511580228806, -0.05898245796561241, -0.016795456409454346, 0.056305158883333206, 0.024857956916093826, -0.02538658119738102, -0.1063234880566597, 0.026648517698049545, 0.024502448737621307, -0.08071450144052505, -0.12649425864219666, 0.051453106105327606, 0.04340458661317825, -0.13367775082588196, -0.015499441884458065, 0.05286165699362755, -0.060711782425642014, 0.033368825912475586, 0.05550144985318184, 0.03024120442569256, 0.11147114634513855, -0.010160478763282299, 0.024509599432349205, 0.06253526359796524, 0.18699023127555847, -0.040643177926540375, 0.11431962996721268, 0.18727125227451324, -0.03192752227187157, 0.1194232925772667, 0.11937419325113297, 0.028797267004847527, -0.049014315009117126, 0.04114986211061478, -0.0008355591562576592, -0.04458879306912422, -0.10618263483047485, -0.04080016911029816, -0.01604492776095867, 0.04535007104277611, 0.09368547797203064, 0.04866218566894531, -0.04582248628139496, 0.0037305953446775675, -0.03777293115854263, -0.022211747244000435, 0.04007827118039131, 0.09904460608959198, -0.008077402599155903, -0.0315714105963707, 0.004549143835902214, -0.011970293708145618, 0.048352960497140884, 0.07191203534603119, 0.11751607060432434, 0.06747867912054062, -0.09643132984638214, 0.11448433995246887, 0.0389794185757637, 0.11041456460952759, 0.0324346087872982, 0.04714079201221466, -0.03517452999949455, 0.05336805060505867, 0.01711718924343586, -0.11136262863874435, 0.015545422211289406, 0.09028099477291107, -0.11122447997331619, 0.01980244554579258, -0.0065583642572164536, 0.09031934291124344, 0.02908378839492798, 0.18630105257034302, 0.021703485399484634, -0.17161989212036133, -0.0002706894592847675, 0.029718399047851562, -0.013328706845641136, -0.08957367390394211, -0.003647462697699666, 0.08437462151050568, -0.12248411029577255, 0.004370574373751879, -0.045610375702381134, 0.05915939062833786, -0.054236434400081635, 0.00173034833278507, 0.09808440506458282, 0.1616477221250534, -0.025058001279830933, 0.10008138418197632, -0.09307120740413666, 0.032082632184028625, 0.018401943147182465, 0.04025441035628319, -0.034390728920698166, 0.023131076246500015, 0.020210202783346176, 0.0824199989438057, 0.12299918383359909, 0.025924162939190865, -0.13950055837631226, -0.059656426310539246, -0.11787029355764389, 0.0344771109521389, 0.030782055109739304, -0.09119714051485062, 0.03387351706624031, -0.02233915962278843, -0.03717200458049774, -0.07615849375724792, -0.04335813596844673, -0.06216789409518242, -0.16486498713493347, 0.06303862482309341, -0.06788752973079681, 0.00663851760327816, -0.04398563504219055, 0.04257851839065552, 0.053914230316877365, 0.088260717689991, -0.0997597873210907, -0.08712474256753922, -0.06378606706857681, -0.06904061138629913, 0.07176955044269562, -0.10366114228963852, -0.0031624624971300364, -0.05030624940991402, 0.16349980235099792, -0.02857639454305172, -0.10962142050266266, 0.035231973975896835, -0.0722479596734047, -0.1006646528840065, -0.025179393589496613, 0.14693227410316467, 0.08662121742963791, 0.0022021192125976086, -0.023878132924437523, 0.0088118277490139, -0.06368156522512436, -0.09386700391769409, -0.05237999185919762, 0.22239437699317932, 0.04788746312260628, -0.013424564152956009, -0.08746637403964996, 0.03972172737121582, -0.06882132589817047, 0.06628553569316864, 0.050583913922309875, 0.18221481144428253, -0.06446132808923721, 0.16217724978923798, 0.1570759117603302, -0.08797284215688705, -0.20304632186889648, -0.03759972006082535, -0.0049448153004050255, 0.021825095638632774, 0.040039438754320145, -0.14832447469234467, 0.13153040409088135, 0.055965375155210495, -0.014148224145174026, 0.08606337010860443, -0.2869226038455963, -0.1259234994649887, 0.037370286881923676, 0.032138045877218246, -0.12093492597341537, -0.0582604855298996, -0.025776075199246407, -0.0828593373298645, -0.08539882302284241, 0.04070904850959778, -0.24039560556411743, 0.0657910481095314, -0.023638702929019928, 0.025836221873760223, 0.03429924696683884, -0.055315203964710236, 0.10814188420772552, -0.046700891107320786, 0.08097974956035614, -0.06652141362428665, 0.06978332251310349, 0.0025417532306164503, -0.1414594054222107, 0.13280871510505676, -0.09234577417373657, 0.07335034012794495, 0.02610470913350582, -0.032847750931978226, -0.030167028307914734, 0.07685300707817078, -0.0658690333366394, -0.06158328056335449, -0.03427868336439133, 0.07337386161088943, 0.05319288372993469, -0.023183155804872513, -0.15217716991901398, -0.06935431063175201, -0.048624761402606964, 0.2341158390045166, 0.08554722368717194, -0.017059583216905594, -0.12334012240171432, 0.014214010909199715, 0.03104095719754696, 0.07030736654996872, -0.06743015348911285, 0.061472419649362564, 0.05627281963825226, 0.04232381284236908, 0.06943628937005997, -0.010834919288754463, -0.12244473397731781, -0.02486373297870159, 0.015217304229736328, -0.15699705481529236, -0.12393658608198166, -0.061835143715143204, 0.12144862115383148, -0.06654965132474899, -0.017242347821593285, 0.12312258034944534, -0.00567018473520875, -0.006927861832082272, 0.04141845181584358, 0.04529670998454094, -0.03513188287615776, 0.0727730393409729, -0.0023503454867750406, 0.025154823437333107, -0.05377427488565445, 0.09494844079017639, 0.031647827476263046, -0.004094212781637907, -0.015141288749873638, 0.17967036366462708, -0.1057298555970192, -0.0878281444311142, -0.08141094446182251, 0.07060742378234863, 0.054128218442201614, -0.08000057935714722, 0.014548661187291145, 0.0375506654381752, -0.07826831191778183, 0.06414734572172165, 0.04041033610701561, 0.006072414573282003, -0.014092561788856983, -0.003366021905094385, -0.05293310433626175, 0.13945819437503815, 0.04757511988282204, 0.04007589444518089, -0.07079866528511047, 0.09281378239393234, 0.03844771161675453, 0.06944914162158966, -0.0192156583070755, -0.07903852313756943, -0.07300551980733871, -0.0028076714370399714, -0.1022825688123703, 0.04197099059820175, -0.04781552404165268, -0.0029487484134733677, 0.0064585027284920216, 0.039049845188856125, 0.035115841776132584, 0.04616932198405266, -0.012590421363711357, -0.030764523893594742, -0.0815097913146019, 0.0383886881172657, -0.09603416919708252, 0.012982298620045185, 0.07887455821037292, -0.07155821472406387, 0.09129993617534637, 0.01741594821214676, -0.07947053760290146, 0.015421725809574127, -0.07495639473199844, 0.024840759113430977, -0.07224218547344208, 0.048751577734947205, 0.03482348844408989, -0.1542588472366333, -0.023210233077406883, -0.05239832401275635, -0.045516230165958405, -0.05307476967573166, 0.10220516473054886, -0.09353604912757874, 0.16218966245651245, 0.012325151823461056, -0.056886378675699234, -0.10017272084951401, -0.011243659071624279, -0.02542976848781109, 0.07093972712755203, 0.037811409682035446, -0.06216159462928772, 0.09330673515796661, -0.11116297543048859, -0.04420582577586174, 0.11581151187419891, 0.036827780306339264, -0.03548157960176468, -0.07660969346761703, 0.03451794758439064, -0.05552995949983597, 0.11234579235315323, 0.029672713950276375, 0.01659421995282173, -0.007041299715638161, -0.044863782823085785, -0.021683407947421074, 0.016955824568867683, -0.036988515406847, -0.09032002836465836, 0.06257972866296768, 0.007673958316445351, 0.010865354910492897, -0.055815789848566055, -0.008988250978291035, 0.10013127326965332, 0.04930004104971886, 0.08698444813489914, 0.0110687967389822, 0.028828412294387817, 0.02689284458756447, -0.14804866909980774, 0.009286620654165745, -0.037264250218868256, 0.05587027221918106, -0.11517000943422318, 0.06703627854585648, 0.120355024933815, -0.12627974152565002, 0.06740408390760422, 0.039504870772361755, -0.10204955190420151, -0.14761210978031158, -0.27984756231307983, -0.017825162038207054, -0.08367306739091873, -0.03823844715952873, -0.049092184752225876, 0.052446767687797546, -0.08684691041707993, 0.023384282365441322, -0.025443261489272118, 0.16125428676605225, -0.044465187937021255, -0.06675519049167633, -0.07335072010755539, 0.032715871930122375, 0.061983298510313034, 0.09565708786249161, 0.047480665147304535, 0.022763922810554504, 0.020985450595617294, 0.010381132364273071, 0.050717875361442566, 0.10874759405851364, 0.0021299547515809536, -0.047435298562049866, -0.01730099506676197, -0.010605734772980213, 0.008072678931057453, -0.04773077368736267, 0.08550246059894562, 0.09557093679904938, -0.0725894644856453, -0.03640012815594673, 0.1593875288963318, -0.07894419878721237, -0.10443352162837982, -0.12196162343025208, 0.27581873536109924, -0.08609682321548462, -0.013316744938492775, -0.007743046618998051, -0.09890937805175781, -0.045923590660095215, 0.18728795647621155, 0.166219100356102, -0.04359117150306702, 0.049514878541231155, -0.009079660288989544, 0.017428409308195114, -0.06101754307746887, 0.074677474796772, 0.048573099076747894, 0.19733400642871857, -0.0357818678021431, 0.07623099535703659, 0.017599163576960564, -0.053047556430101395, -0.14523395895957947, 0.03712725639343262, -0.09411325305700302, 0.019424080848693848, -0.006762867793440819, 0.053080979734659195, -0.029323864728212357, -0.22161069512367249, -0.035006701946258545, 0.07119831442832947, -0.05649492144584656, 0.03169206157326698, 0.022959373891353607, -0.04441513493657112, 0.045400314033031464, -0.040586646646261215, -0.022539708763360977, 0.23096226155757904, -0.012305007316172123, -0.060960594564676285, -0.03406694531440735, 0.08848007768392563, -0.08970090001821518, 0.14024437963962555, 0.01789834164083004, 0.028461502864956856, 0.05203433707356453, 0.004104297142475843, -0.07445421814918518, 0.06558804214000702, 0.03542299196124077, -0.12915557622909546, -0.028101634234189987, 0.1356048732995987, 0.012820009142160416, 0.004372606985270977, 0.0337335504591465, -0.0400085523724556, 0.0020790747366845608, -0.07823818922042847, 0.04329817369580269, -0.13793586194515228, 0.06881414353847504, -0.06333695352077484, 0.14489413797855377, 0.15732131898403168, -0.01459522545337677, 0.003891671309247613, -0.03214459493756294, 0.021356185898184776, 0.0014611221849918365, -0.03490697219967842, 0.016994230449199677, -0.08531537652015686, 0.04047037661075592, 0.02947978861629963, 0.04495260864496231, -0.14632098376750946, -0.0724063590168953, 0.021890979260206223, -0.021044112741947174, 0.05450300872325897, 0.11902664601802826, 0.021243447437882423, 0.05827771872282028, -0.01673201099038124, -0.045240506529808044, 0.004287718329578638, 0.07524386793375015, -0.15941523015499115, -0.06728441268205643 ]
null
null
peft
## Training procedure The following `bitsandbytes` quantization config was used during training: - load_in_8bit: False - load_in_4bit: True - llm_int8_threshold: 6.0 - llm_int8_skip_modules: None - llm_int8_enable_fp32_cpu_offload: False - llm_int8_has_fp16_weight: False - bnb_4bit_quant_type: nf4 - bnb_4bit_use_double_quant: False - bnb_4bit_compute_dtype: float16 ### Framework versions - PEFT 0.4.0
{"language": ["en"], "library_name": "peft", "tags": ["facebook", "meta", "pytorch", "llama", "llama-2", "text generation", "hosted inference"], "inference": true, "pipeline_tag": "text-generation"}
text-generation
Shri2818/codellama_python_v1
[ "peft", "facebook", "meta", "pytorch", "llama", "llama-2", "text generation", "hosted inference", "text-generation", "en", "region:us" ]
2024-02-10T14:19:46+00:00
[]
[ "en" ]
TAGS #peft #facebook #meta #pytorch #llama #llama-2 #text generation #hosted inference #text-generation #en #region-us
## Training procedure The following 'bitsandbytes' quantization config was used during training: - load_in_8bit: False - load_in_4bit: True - llm_int8_threshold: 6.0 - llm_int8_skip_modules: None - llm_int8_enable_fp32_cpu_offload: False - llm_int8_has_fp16_weight: False - bnb_4bit_quant_type: nf4 - bnb_4bit_use_double_quant: False - bnb_4bit_compute_dtype: float16 ### Framework versions - PEFT 0.4.0
[ "## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: False\n- bnb_4bit_compute_dtype: float16", "### Framework versions\n\n\n- PEFT 0.4.0" ]
[ "TAGS\n#peft #facebook #meta #pytorch #llama #llama-2 #text generation #hosted inference #text-generation #en #region-us \n", "## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: False\n- bnb_4bit_compute_dtype: float16", "### Framework versions\n\n\n- PEFT 0.4.0" ]
[ 39, 154, 11 ]
[ "passage: TAGS\n#peft #facebook #meta #pytorch #llama #llama-2 #text generation #hosted inference #text-generation #en #region-us \n## Training procedure\n\n\nThe following 'bitsandbytes' quantization config was used during training:\n- load_in_8bit: False\n- load_in_4bit: True\n- llm_int8_threshold: 6.0\n- llm_int8_skip_modules: None\n- llm_int8_enable_fp32_cpu_offload: False\n- llm_int8_has_fp16_weight: False\n- bnb_4bit_quant_type: nf4\n- bnb_4bit_use_double_quant: False\n- bnb_4bit_compute_dtype: float16### Framework versions\n\n\n- PEFT 0.4.0" ]
[ -0.06313753128051758, 0.031422924250364304, -0.004081635735929012, 0.13320007920265198, 0.09143409132957458, 0.028025396168231964, 0.12259084731340408, 0.12358935177326202, 0.04571162164211273, 0.09666290879249573, 0.10702767968177795, 0.055812373757362366, 0.08015037328004837, 0.11617910861968994, -0.04076550528407097, -0.022507285699248314, 0.02841256372630596, -0.030765259638428688, 0.0004888082039542496, 0.08403748273849487, 0.051697734743356705, -0.03195032849907875, 0.06806211173534393, -0.08649563789367676, -0.13738393783569336, 0.029363272711634636, 0.00029362126952037215, 0.019799914211034775, 0.027231670916080475, 0.0587727352976799, 0.023047873750329018, 0.015016879886388779, -0.0434269979596138, -0.17926989495754242, -0.0012805084697902203, 0.09813442081212997, -0.022426927462220192, 0.060208361595869064, -0.03547554090619087, 0.039232514798641205, 0.054672859609127045, -0.045501358807086945, 0.00869743898510933, 0.033901870250701904, -0.08411266654729843, -0.10612597316503525, -0.06234414875507355, 0.10051792860031128, -0.004809437785297632, 0.09513598680496216, -0.016590148210525513, 0.13499219715595245, -0.0857648253440857, 0.10579577088356018, 0.11826492100954056, -0.20080895721912384, -0.03721588850021362, 0.10906241834163666, -0.024882692843675613, 0.1197907105088234, -0.05118996649980545, -0.07232683151960373, 0.08706426620483398, 0.0697937160730362, -0.06576098501682281, -0.006706536281853914, -0.08602023124694824, -0.03174540773034096, -0.12086311727762222, -0.04502199590206146, 0.17796993255615234, 0.03707519918680191, -0.03534025326371193, -0.055400550365448, -0.08080174773931503, -0.23453570902347565, 0.03944362327456474, 0.002581317676231265, -0.06793083995580673, 0.052402593195438385, 0.046975892037153244, -0.04807300865650177, -0.02023843675851822, -0.06669726967811584, -0.04480130970478058, 0.08222553879022598, 0.021320346742868423, 0.04216678440570831, -0.017456306144595146, 0.0919651910662651, -0.07989614456892014, -0.06673844158649445, -0.010210562497377396, -0.04054674878716469, 0.040360450744628906, 0.019503748044371605, -0.06081904470920563, 0.09660430252552032, 0.10522188246250153, 0.023776741698384285, -0.1982920914888382, 0.11083240061998367, -0.08290235698223114, 0.06749822199344635, -0.02829466201364994, -0.034645020961761475, -0.1358002871274948, 0.0902990847826004, 0.01645299233496189, 0.14766858518123627, 0.03849612921476364, -0.0414666086435318, -0.05963890627026558, 0.005050214473158121, 0.10774551331996918, 0.01586958020925522, -0.07184194773435593, 0.034422148019075394, -0.1163424625992775, -0.03468022868037224, 0.03749373182654381, -0.09672822058200836, -0.002428452717140317, 0.07505294680595398, -0.07977188378572464, 0.06532541662454605, 0.13577407598495483, -0.044322073459625244, -0.08066292852163315, -0.07695778459310532, -0.08015961945056915, 0.007087273523211479, -0.06593407690525055, -0.13331900537014008, 0.058908987790346146, -0.113113634288311, -0.03242307901382446, -0.04834548756480217, -0.13904793560504913, -0.01713518053293228, -0.0017054600175470114, -0.08457659929990768, 0.022575054317712784, -0.11606284976005554, -0.16807888448238373, -0.03814856335520744, 0.0073984600603580475, 0.029594089835882187, -0.04065316915512085, 0.11315085738897324, 0.0008174625691026449, 0.08096194267272949, -0.15670472383499146, 0.005871668457984924, 0.007513604126870632, 0.07454650849103928, -0.0007062585791572928, 0.18046842515468597, -0.08366327732801437, 0.005911946762353182, -0.07174789160490036, -0.07729645818471909, -0.08779507130384445, -0.010582735762000084, 0.12537553906440735, 0.13506370782852173, -0.18922244012355804, -0.03206130489706993, 0.07797272503376007, -0.04759472608566284, -0.09956485033035278, 0.12707404792308807, -0.048869937658309937, 0.1221052035689354, 0.007251352071762085, 0.09603458642959595, 0.19500353932380676, -0.09775324165821075, 0.0032048507127910852, 0.08590283244848251, 0.051081474870443344, 0.051473867148160934, -0.019739193841814995, 0.04354419931769371, -0.167598158121109, 0.050900477916002274, 0.09622542560100555, 0.0308570247143507, -0.05521732196211815, -0.06843673437833786, -0.034606169909238815, -0.07033175975084305, 0.09202244877815247, 0.018631819635629654, -0.0058913505636155605, -0.0742335096001625, -0.07483134418725967, 0.08368406444787979, 0.11421303451061249, -0.028758693486452103, -0.024996979162096977, -0.10260993987321854, 0.04918750375509262, -0.014615901745855808, 0.02704174630343914, -0.1363355815410614, 0.0019263246795162559, 0.04413851350545883, 0.0434858500957489, 0.016970336437225342, 0.05981970578432083, 0.06393902748823166, 0.021979134529829025, -0.031333230435848236, 0.03521956130862236, -0.018239665776491165, -0.0005013197660446167, -0.07644318789243698, -0.14449705183506012, 0.034925591200590134, -0.011727764271199703, 0.17517897486686707, -0.18926411867141724, 0.049748919904232025, 0.1284715235233307, 0.04422265291213989, -0.04340045526623726, -0.018969327211380005, -0.037365276366472244, 0.07229995727539062, 0.0019274932565167546, -0.03496837243437767, 0.0520392581820488, 0.026690755039453506, -0.08314168453216553, -0.11971042305231094, -0.14589770138263702, 0.02506321482360363, 0.1271435171365738, 0.04087977111339569, -0.0904020294547081, -0.062438055872917175, -0.031865738332271576, -0.022158194333314896, 0.06919295340776443, -0.04683919623494148, 0.09341017156839371, 0.012989088892936707, 0.0826689600944519, -0.10346712172031403, -0.03463006019592285, 0.09220115095376968, 0.0008227091166190803, -0.049525145441293716, 0.14835701882839203, -0.004458849783986807, -0.05923778563737869, 0.08893800526857376, 0.04384998604655266, -0.09329768270254135, 0.11204885691404343, 0.00803843978792429, -0.011774454265832901, -0.09794911742210388, 0.18733185529708862, -0.0004997991491109133, 0.09182415157556534, -0.15393324196338654, 0.09626488387584686, -0.007463627494871616, -0.003171020420268178, 0.07277830690145493, -0.16292273998260498, -0.009424789808690548, -0.05324864014983177, -0.08254865556955338, -0.028398552909493446, 0.00031818225397728384, 0.05448164790868759, 0.08720137923955917, -0.013278959318995476, 0.0350334569811821, 0.11339694261550903, -0.038466986268758774, -0.1099046841263771, 0.19421981275081635, -0.18540523946285248, -0.26088252663612366, -0.16991372406482697, 0.02237694337964058, -0.1729642152786255, -0.034239452332258224, -0.022813310846686363, -0.07853817939758301, 0.030516086146235466, -0.11321917176246643, -0.009510346688330173, -0.018746627494692802, -0.005168781150132418, 0.02794674225151539, -0.004265869967639446, 0.16662970185279846, -0.07985468208789825, -0.008802066557109356, 0.03369264304637909, -0.030577247962355614, 0.07845200598239899, -0.06903935968875885, -0.058830246329307556, 0.1114753782749176, 0.0019811526872217655, 0.029919691383838654, -0.008401782251894474, 0.3196606934070587, 0.01410562265664339, 0.021219197660684586, 0.13756997883319855, -0.0018064682371914387, 0.07857287675142288, 0.09417056292295456, 0.012615565210580826, -0.13348017632961273, 0.07403803616762161, 0.06088367477059364, -0.06233421713113785, -0.15299636125564575, -0.06851188093423843, -0.05228499323129654, 0.03935115784406662, 0.08536573499441147, 0.06976036727428436, 0.06935387849807739, 0.06319347769021988, 0.004532660823315382, 0.09136547148227692, 0.040158167481422424, 0.0005665582139045, 0.1456502228975296, -0.04865717142820358, 0.06243084743618965, -0.008683349937200546, 0.022745544090867043, 0.08143231272697449, 0.08696816861629486, 0.0240408293902874, -0.12951788306236267, -0.0782460942864418, 0.08916130661964417, 0.20103827118873596, -0.032648731023073196, 0.06349831819534302, -0.057704318314790726, 0.014566016383469105, -0.0016907269600778818, -0.07440412789583206, -0.08594118803739548, 0.03424784913659096, -0.03232721611857414, 0.10212033241987228, -0.061888307332992554, -0.032817136496305466, 0.07930319756269455, 0.1257726401090622, 0.13294629752635956, -0.27269452810287476, -0.1086101159453392, -0.01149686798453331, 0.0929090678691864, -0.09534671157598495, 0.04688923433423042, 0.1987932175397873, -0.017821265384554863, -0.006127046886831522, -0.016627002507448196, 0.041908297687768936, -0.012436307966709137, 0.040544234216213226, 0.060312420129776, 0.11666480451822281, -0.007581460755318403, 0.08997570723295212, -0.31924691796302795, 0.010781237855553627, 0.038541555404663086, 0.0351630263030529, -0.06231413036584854, 0.0011570993810892105, -0.025267675518989563, -0.06980271637439728, 0.08340604603290558, 0.0069478657096624374, 0.21192696690559387, -0.2049032747745514, -0.07925776392221451, -0.0032882140949368477, 0.07834385335445404, 0.036413151770830154, 0.03776587173342705, 0.041097063571214676, 0.06559941172599792, 0.04550975188612938, 0.020708143711090088, -0.06364332139492035, -0.10345251113176346, 0.011063124053180218, 0.11888911575078964, -0.11206860840320587, -0.050112683326005936, -0.07658471912145615, -0.053716231137514114, 0.11740560829639435, -0.19034934043884277, -0.0964953750371933, -0.06731154024600983, 0.07068716734647751, 0.10627669095993042, -0.012751415371894836, -0.04041330888867378, -0.023528357967734337, 0.03623896837234497, -0.03741847723722458, -0.10319926589727402, 0.10089772194623947, -0.029685843735933304, -0.1688132882118225, 0.0041555375792086124, 0.16523787379264832, 0.1103443130850792, -0.018169617280364037, -0.07034869492053986, -0.04520376771688461, 0.001756183453835547, -0.1575261950492859, 0.0032037983182817698, 0.13188450038433075, -0.03205389156937599, 0.1079258918762207, -0.08688750863075256, 0.16034552454948425, -0.07369223982095718, 0.09331263601779938, 0.06092184782028198, 0.3005684018135071, -0.06931120902299881, 0.04711812734603882, 0.05283704772591591, -0.055996980518102646, -0.23077493906021118, 0.037229590117931366, 0.050904903560876846, 0.03015361726284027, -0.046574074774980545, -0.15388023853302002, 0.02993069589138031, 0.07255632430315018, 0.03504880517721176, 0.1908690631389618, -0.32713645696640015, -0.07980789989233017, 0.06617434322834015, 0.07220354676246643, 0.17880025506019592, -0.06575024127960205, -0.007412200793623924, -0.017477499321103096, 0.01944931410253048, 0.18641065061092377, -0.13598576188087463, 0.09764890372753143, -0.018341323360800743, 0.07993894070386887, 0.028181301429867744, -0.01612779311835766, 0.13917674124240875, 0.005728818941861391, 0.08625876158475876, -0.01225955318659544, -0.013210039585828781, 0.036354776471853256, -0.05622373893857002, 0.07700810581445694, -0.11182162910699844, 0.07927799969911575, -0.09084007889032364, -0.0053792246617376804, -0.0927853211760521, 0.030189618468284607, -0.09212891757488251, -0.022334739565849304, -0.08550553768873215, 0.04851093515753746, 0.03420400992035866, -0.003050711937248707, -0.06221979856491089, 0.0082289669662714, 0.11333037912845612, 0.42077741026878357, 0.0059949676506221294, -0.01470804214477539, 0.06907812505960464, 0.09688948094844818, -0.02011754736304283, 0.10751297324895859, -0.13591767847537994, 0.03682570531964302, 0.14580926299095154, 0.012961364351212978, 0.14561301469802856, 0.07836798578500748, -0.12597540020942688, -0.017794061452150345, 0.02281823381781578, -0.18545936048030853, -0.06642038375139236, -0.04404403269290924, 0.0041221813298761845, -0.08920864760875702, -0.027624279260635376, 0.1541232317686081, -0.059286605566740036, 0.049026183784008026, 0.030134471133351326, 0.049652233719825745, -0.11152607202529907, 0.10668306797742844, 0.04402302950620651, 0.061806581914424896, -0.10720789432525635, 0.10915815085172653, 0.020239006727933884, -0.02591177076101303, 0.08774526417255402, 0.03980983421206474, -0.09659069031476974, -0.010927342809736729, -0.05862108990550041, -0.0013829153031110764, 0.09189644455909729, -0.05613508075475693, -0.06716785579919815, -0.140110045671463, 0.0284115020185709, 0.1632298082113266, 0.047373756766319275, 0.10350708663463593, -0.018196024000644684, -0.023926299065351486, -0.10547972470521927, 0.09934101998806, -0.00027734029572457075, -0.018140826374292374, -0.11058969795703888, 0.0964038297533989, -0.00608514528721571, 0.0780668705701828, -0.03384893015027046, -0.026877257972955704, -0.21380886435508728, 0.003738815663382411, -0.10982241481542587, 0.035144951194524765, 0.020474065095186234, 0.032012131065130234, 0.016682447865605354, 0.04179316386580467, -0.02843630127608776, 0.027183562517166138, -0.03377263993024826, -0.03728468343615532, 0.04746271297335625, 0.026958880946040154, -0.07710638642311096, -0.06399284303188324, 0.020448138937354088, -0.1026141569018364, 0.04124990105628967, 0.056712884455919266, -0.07419365644454956, 0.04615894332528114, 0.01581462472677231, 0.03277468681335449, 0.07969886064529419, 0.06697800010442734, 0.02100461535155773, -0.07601286470890045, 0.03406289592385292, -0.006750107742846012, -0.012993043288588524, 0.061041444540023804, 0.1523759961128235, -0.06016376614570618, -0.05326876789331436, -0.05536137521266937, -0.013075211085379124, -0.044296130537986755, 0.044797852635383606, 0.11411592364311218, 0.12438593804836273, 0.08135621249675751, -0.11167003214359283, 0.016745254397392273, -0.15813571214675903, -0.06659543514251709, 0.05822022259235382, -0.0313635915517807, 0.020383687689900398, -0.01435889396816492, 0.0938861221075058, -0.004227441735565662, 0.152144655585289, -0.04745032638311386, -0.08445427566766739, -0.03800470381975174, -0.16931618750095367, -0.11296409368515015, -0.0018662321381270885, 0.1601482331752777, 0.0038346187211573124, -0.02070031873881817, -0.050176069140434265, -0.01196388341486454, 0.08202921599149704, 0.13546456396579742, 0.07921264320611954, 0.046993058174848557, -0.08449038118124008, 0.06562739610671997, 0.07026361674070358, -0.04843490570783615, 0.1419423669576645, 0.2519640326499939, -0.048654619604349136, 0.04557392746210098, -0.10011652112007141, 0.06562776863574982, 0.031206605955958366, -0.1205977201461792, 0.027496229857206345, -0.0424475260078907, -0.1487702876329422, -0.11270101368427277, -0.02474246546626091, -0.05280870944261551, -0.15221397578716278, -0.02397877164185047, -0.1188916489481926, -0.06012137606739998, 0.0804789736866951, 0.04410099610686302, -0.04407592490315437, 0.14393287897109985, -0.09261555969715118, -0.007589196786284447, 0.02134270779788494, 0.0033969541545957327, 0.007756740786135197, -0.049228914082050323, -0.0672747865319252, 0.1536281555891037, 0.03645167499780655, 0.06006469205021858, 0.03680848702788353, 0.1029280349612236, 0.028285417705774307, -0.040292639285326004, -0.05152774229645729, 0.0024143538903445005, 0.007943824864923954, -0.046523068100214005, 0.14181481301784515, 0.07155030965805054, -0.07427989691495895, -0.07036104798316956, 0.0025220736861228943, -0.08269575238227844, -0.019025035202503204, -0.15371379256248474, 0.2614573538303375, -0.06414175033569336, 0.0738544836640358, -0.024229135364294052, -0.051919348537921906, -0.10283263772726059, 0.1705034375190735, 0.14915941655635834, -0.160720095038414, 0.010248446837067604, 0.0488264337182045, -0.0008472857298329473, -0.0403839647769928, 0.09766312688589096, 0.09836336970329285, 0.028358280658721924, 0.012884842231869698, 0.005106019321829081, -0.040191199630498886, -0.00623876228928566, -0.014373360201716423, 0.005843305494636297, 0.007279639132320881, 0.02967672049999237, -0.11056025326251984, 0.002991014625877142, -0.11477119475603104, -0.0884733721613884, 0.14048531651496887, -0.14326494932174683, -0.08779257535934448, -0.036182306706905365, -0.12202563136816025, -0.05615317448973656, 0.02109440788626671, -0.10585404187440872, 0.05746928229928017, 0.10778681188821793, -0.06445957720279694, -0.02013661153614521, -0.06791436672210693, -0.01711254008114338, -0.001661138841882348, 0.06894427537918091, -0.016620351001620293, 0.0693865641951561, 0.09820857644081116, -0.0228154044598341, -0.04417524114251137, 0.07738221436738968, 0.017145676538348198, -0.008020620793104172, -0.11613915860652924, 0.03409337252378464, 0.00310850259847939, 0.11913721263408661, 0.0641433522105217, -0.03484978526830673, -0.040114253759384155, -0.11018679291009903, -0.05152571201324463, -0.1163829043507576, -0.056314412504434586, -0.06693747639656067, 0.08620401471853256, 0.1969359815120697, -0.06586840748786926, 0.028561126440763474, -0.04193136841058731, 0.017931818962097168, -0.05703091621398926, 0.02895219810307026, 0.03167795389890671, -0.16051499545574188, 0.033007558435201645, -0.007946554571390152, 0.02299395762383938, -0.33737146854400635, -0.0034356347750872374, -0.018340861424803734, -0.015276017598807812, -0.04218820109963417, 0.11610899120569229, -0.04006936401128769, 0.05125753581523895, -0.06390020996332169, -0.16218219697475433, -0.054713308811187744, 0.15624740719795227, -0.004003262612968683, -0.07809187471866608 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # SpeechT5 TTS Dutch This model is a fine-tuned version of [microsoft/speecht5_tts](https://huggingface.co/microsoft/speecht5_tts) on the VoxPopuli dataset. It achieves the following results on the evaluation set: - Loss: 0.4555 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 16 - eval_batch_size: 8 - seed: 42 - gradient_accumulation_steps: 2 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 500 - training_steps: 4000 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 0.5202 | 4.3 | 1000 | 0.4754 | | 0.4983 | 8.6 | 2000 | 0.4625 | | 0.4972 | 12.9 | 3000 | 0.4574 | | 0.4953 | 17.2 | 4000 | 0.4555 | ### Framework versions - Transformers 4.38.0.dev0 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.1
{"language": ["nl"], "license": "mit", "tags": ["generated_from_trainer"], "datasets": ["facebook/voxpopuli"], "base_model": "microsoft/speecht5_tts", "model-index": [{"name": "SpeechT5 TTS Dutch", "results": []}]}
text-to-audio
IoanRazvan/speecht5_tts_voxpopuli_nl
[ "transformers", "tensorboard", "safetensors", "speecht5", "text-to-audio", "generated_from_trainer", "nl", "dataset:facebook/voxpopuli", "base_model:microsoft/speecht5_tts", "license:mit", "endpoints_compatible", "region:us" ]
2024-02-10T14:23:01+00:00
[]
[ "nl" ]
TAGS #transformers #tensorboard #safetensors #speecht5 #text-to-audio #generated_from_trainer #nl #dataset-facebook/voxpopuli #base_model-microsoft/speecht5_tts #license-mit #endpoints_compatible #region-us
SpeechT5 TTS Dutch ================== This model is a fine-tuned version of microsoft/speecht5\_tts on the VoxPopuli dataset. It achieves the following results on the evaluation set: * Loss: 0.4555 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 1e-05 * train\_batch\_size: 16 * eval\_batch\_size: 8 * seed: 42 * gradient\_accumulation\_steps: 2 * total\_train\_batch\_size: 32 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 500 * training\_steps: 4000 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.38.0.dev0 * Pytorch 2.1.0+cu121 * Datasets 2.17.0 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 4000\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #tensorboard #safetensors #speecht5 #text-to-audio #generated_from_trainer #nl #dataset-facebook/voxpopuli #base_model-microsoft/speecht5_tts #license-mit #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 4000\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ 76, 158, 4, 38 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #speecht5 #text-to-audio #generated_from_trainer #nl #dataset-facebook/voxpopuli #base_model-microsoft/speecht5_tts #license-mit #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 8\n* seed: 42\n* gradient\\_accumulation\\_steps: 2\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 500\n* training\\_steps: 4000\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.38.0.dev0\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ -0.1327894628047943, 0.09192629903554916, -0.003287276718765497, 0.05167907848954201, 0.09680274873971939, 0.007510046940296888, 0.09464535862207413, 0.1500614732503891, -0.08357959240674973, 0.10749813169240952, 0.09118273109197617, 0.05934196710586548, 0.06848747283220291, 0.2084183245897293, -0.030840812250971794, -0.2953454852104187, 0.015554244630038738, -0.014800330623984337, -0.1332014948129654, 0.11526311933994293, 0.08815094828605652, -0.10578068345785141, 0.03319820016622543, -0.011071419343352318, -0.10036339610815048, -0.012892615981400013, -0.02126016467809677, -0.048483263701200485, 0.10495194047689438, 0.045359812676906586, 0.06564116477966309, 0.061032164841890335, 0.07873726636171341, -0.2512328028678894, 0.016572697088122368, 0.0768074095249176, 0.024631746113300323, 0.07156869024038315, 0.11439871042966843, -0.020052755251526833, 0.11459000408649445, -0.09242357313632965, 0.06270255148410797, 0.04156098887324333, -0.11333472281694412, -0.31441599130630493, -0.10693913698196411, 0.03337389603257179, 0.1511574685573578, 0.06352782994508743, -0.03901061788201332, 0.07374851405620575, -0.05434688925743103, 0.0978986993432045, 0.2224271148443222, -0.25007596611976624, -0.07253603637218475, 0.00006709424633299932, 0.09447517991065979, 0.07034821808338165, -0.1064598485827446, -0.003745822701603174, 0.03350454568862915, 0.0233139730989933, 0.12233830243349075, -0.00259271333925426, 0.06649583578109741, -0.010373394005000591, -0.14322733879089355, -0.04007188603281975, 0.09616579860448837, 0.08461958169937134, -0.0295262448489666, -0.13002975285053253, -0.021409714594483376, -0.19472576677799225, -0.04209543392062187, 0.0035809532273560762, 0.02318476513028145, -0.04212711751461029, -0.11733438074588776, 0.00860190112143755, -0.053072717040777206, -0.07301513850688934, 0.04600367695093155, 0.10277162492275238, 0.03762294724583626, -0.036617059260606766, 0.018396718427538872, 0.10680738836526871, 0.019731009379029274, -0.1554037481546402, 0.010677455924451351, 0.020323920994997025, -0.11719408631324768, -0.042412109673023224, -0.018691781908273697, -0.0528094545006752, 0.013435115106403828, 0.15691371262073517, -0.046815793961286545, 0.08998289704322815, 0.019602103158831596, 0.029384495690464973, -0.059720586985349655, 0.11251933127641678, -0.06286316365003586, -0.10779885202646255, -0.04327467456459999, 0.11152458190917969, 0.004759605973958969, -0.012486180290579796, -0.07379204034805298, 0.026156019419431686, 0.07901530712842941, 0.05253210291266441, 0.0037796793039888144, 0.01140732504427433, -0.09241568297147751, -0.02342921867966652, 0.03300715982913971, -0.08893435448408127, 0.06282391399145126, 0.02054223231971264, -0.0417972169816494, -0.0421263612806797, -0.007367201615124941, 0.023253340274095535, 0.0027129335794597864, 0.14394529163837433, -0.04044768959283829, -0.018591754138469696, -0.08499306440353394, -0.10626484453678131, 0.044980600476264954, -0.05637199804186821, -0.000563328096177429, -0.05070948600769043, -0.09497864544391632, -0.06639652699232101, 0.0760369747877121, -0.05467834323644638, -0.059373658150434494, -0.05761536955833435, -0.048535365611314774, 0.04998558387160301, -0.03722655028104782, 0.1778663545846939, -0.06921988725662231, 0.1145617663860321, -0.00234503042884171, 0.08050629496574402, 0.06536763161420822, 0.06796947121620178, -0.0289789829403162, 0.06635643541812897, -0.22012171149253845, 0.07988527417182922, -0.09936441481113434, 0.04160112887620926, -0.1469140499830246, -0.1033390685915947, -0.016881896182894707, 0.006221994757652283, 0.09814262390136719, 0.10347528755664825, -0.18534968793392181, -0.10559619218111038, 0.1661493182182312, -0.08151747286319733, -0.09643194079399109, 0.14046435058116913, -0.02315676026046276, -0.006585312075912952, 0.03350663185119629, 0.1866542100906372, 0.11132040619850159, -0.12502916157245636, 0.010422559455037117, -0.05724996700882912, 0.08837077766656876, 0.042090218514204025, 0.08695206046104431, -0.03272045776247978, 0.05569179728627205, -0.018772972747683525, -0.00208646384999156, 0.07514186948537827, -0.07870422303676605, -0.06656299531459808, -0.002236412838101387, -0.07389139384031296, 0.038295306265354156, 0.046519685536623, 0.007205320987850428, -0.10296064615249634, -0.11981777846813202, 0.039656441658735275, 0.09691435098648071, -0.0809636116027832, 0.033159513026475906, -0.061294060200452805, 0.04244570806622505, -0.02501020021736622, -0.014315921813249588, -0.16989344358444214, 0.0051124622114002705, 0.027250755578279495, -0.06237694248557091, 0.010537576861679554, -0.0280244592577219, 0.07932045310735703, 0.049530863761901855, -0.08357242494821548, -0.07310637831687927, -0.03461700305342674, 0.011329757049679756, -0.08061698079109192, -0.24817289412021637, -0.06524593383073807, -0.04540460184216499, 0.1315501630306244, -0.19725476205348969, 0.00960507057607174, 0.06382060796022415, 0.1518823504447937, 0.06704395264387131, -0.052119724452495575, 0.02720646932721138, 0.08906284719705582, 0.0016408558003604412, -0.0858079120516777, 0.03485488519072533, 0.01190838310867548, -0.14122529327869415, 0.00993182323873043, -0.13249064981937408, 0.1059446781873703, 0.08732575923204422, 0.039491474628448486, -0.11341941356658936, -0.07642433047294617, -0.060032859444618225, -0.06307832896709442, -0.03851274400949478, 0.01803041435778141, 0.14442868530750275, 0.040083326399326324, 0.10554319620132446, -0.07457025349140167, -0.04872272163629532, 0.04237326607108116, 0.003676666645333171, -0.010933306999504566, 0.1443948894739151, 0.0266423299908638, -0.09126129746437073, 0.10950371623039246, 0.12250889092683792, -0.029079433530569077, 0.16793055832386017, -0.08249598741531372, -0.10476063936948776, -0.04175931215286255, 0.03894556313753128, 0.03780926764011383, 0.12289755046367645, -0.09607302397489548, 0.006232196930795908, 0.018183525651693344, 0.03415972366929054, 0.005697678774595261, -0.1798826903104782, -0.007318553980439901, 0.048031285405159, -0.060513146221637726, -0.02253560721874237, -0.014736422337591648, -0.006461129989475012, 0.0770539790391922, 0.01341947540640831, -0.029610009863972664, 0.001827568281441927, -0.017969302833080292, -0.0821593627333641, 0.1654110699892044, -0.109244704246521, -0.15815973281860352, -0.12101946026086807, -0.035881027579307556, 0.01166884507983923, -0.01814153417944908, 0.05808992683887482, -0.09245279431343079, -0.03284205123782158, -0.05850899592041969, 0.022607076913118362, -0.03140290826559067, 0.022754766047000885, -0.02055778168141842, 0.02924993634223938, 0.07783396542072296, -0.07845310866832733, 0.030387435108423233, 0.0021081247832626104, 0.004627085290849209, 0.027614032849669456, 0.020541738718748093, 0.08314499258995056, 0.14502394199371338, 0.04007427394390106, 0.022370966151356697, -0.057034995406866074, 0.15699799358844757, -0.1465083211660385, 0.022511839866638184, 0.11603621393442154, -0.009617246687412262, 0.05042586848139763, 0.1714356243610382, 0.043213535100221634, -0.0875476822257042, 0.02774064987897873, 0.02872610278427601, -0.02494223602116108, -0.22809182107448578, -0.02802816592156887, -0.06529077887535095, 0.012574049644172192, 0.10569746047258377, 0.03048735484480858, -0.016099754720926285, 0.026571154594421387, -0.02291303128004074, -0.00924578309059143, 0.03308624029159546, 0.06506326049566269, 0.03698114678263664, 0.03300022333860397, 0.11052219569683075, -0.023348426446318626, -0.007221091538667679, 0.053193073719739914, 0.010444536805152893, 0.23773206770420074, -0.0009077327558770776, 0.17044049501419067, 0.05027660354971886, 0.1342945098876953, 0.018666313961148262, 0.03503766283392906, 0.015484988689422607, -0.022175420075654984, 0.006030789576470852, -0.05519397184252739, -0.000804101990070194, 0.040227651596069336, 0.08165593445301056, 0.005385730881243944, -0.12198107689619064, -0.006327870301902294, 0.023508010432124138, 0.3167160451412201, 0.08676594495773315, -0.2726042568683624, -0.087816022336483, 0.020065978169441223, -0.06380754709243774, -0.0503193661570549, 0.01584283821284771, 0.15586358308792114, -0.08864948898553848, 0.0948561355471611, -0.08018554002046585, 0.09619920700788498, -0.06540088355541229, -0.003918450325727463, 0.08697383850812912, 0.08652181923389435, -0.02688813954591751, 0.040202509611845016, -0.26192495226860046, 0.2971121072769165, 0.008747745305299759, 0.08007998019456863, -0.018053920939564705, 0.03669567406177521, 0.027744080871343613, 0.00040223702671937644, 0.10351874679327011, -0.006673128344118595, -0.18932436406612396, -0.16720373928546906, -0.10083949565887451, -0.009256771765649319, 0.12584340572357178, -0.04586399346590042, 0.0944620817899704, -0.01535861287266016, -0.03514430299401283, 0.03891455754637718, -0.08432409167289734, -0.096367746591568, -0.10594833642244339, 0.01644979789853096, 0.010958408936858177, 0.06864533573389053, -0.10556276142597198, -0.09500377625226974, -0.05205577611923218, 0.17233961820602417, -0.06342688202857971, -0.024572771042585373, -0.15346336364746094, 0.08097487688064575, 0.1509539932012558, -0.058576084673404694, 0.06802580505609512, 0.023445188999176025, 0.11352722346782684, -0.006025657057762146, -0.004430859815329313, 0.14328347146511078, -0.0717911347746849, -0.21371354162693024, -0.08159396052360535, 0.1783803254365921, 0.03883892297744751, 0.0693548172712326, -0.034446388483047485, 0.043912772089242935, 0.003966610878705978, -0.06280411779880524, 0.08226863294839859, 0.005898861214518547, 0.03826063126325607, 0.03304288536310196, -0.012991897761821747, -0.012972685508430004, -0.033194657415151596, -0.09195052832365036, 0.10616417974233627, 0.3109546899795532, -0.09647217392921448, 0.08407735079526901, 0.06593475490808487, -0.043008167296648026, -0.16565068066120148, 0.03827337175607681, 0.12429310381412506, 0.05155028775334358, 0.047361649572849274, -0.19590020179748535, 0.002377841155976057, 0.08040277659893036, -0.026055583730340004, 0.079665906727314, -0.31438738107681274, -0.13577435910701752, 0.06190260872244835, 0.0746469795703888, -0.06178363785147667, -0.14497911930084229, -0.06404860317707062, -0.016203708946704865, -0.1022787019610405, 0.01740317977964878, -0.05383140593767166, 0.12414925545454025, 0.017235510051250458, 0.01418780442327261, 0.02205977961421013, -0.04661617800593376, 0.12756535410881042, -0.018335193395614624, 0.07035025954246521, -0.011853576637804508, 0.052598610520362854, -0.014924845658242702, -0.0769255980849266, -0.015985744073987007, -0.09139217436313629, 0.012175933457911015, -0.11169618368148804, -0.026209210976958275, -0.06859632581472397, 0.02737264148890972, -0.04933866485953331, -0.04003794863820076, -0.037363216280937195, 0.06949689239263535, 0.06269121915102005, -0.022201847285032272, 0.11782879382371902, -0.059175197035074234, 0.15046241879463196, 0.12775593996047974, 0.102311871945858, -0.0034446557983756065, -0.10553425550460815, 0.0034819929860532284, -0.038086626678705215, 0.04496782645583153, -0.1411060094833374, 0.03721672669053078, 0.13791713118553162, 0.03803941607475281, 0.1420735865831375, 0.03711633011698723, -0.07755286246538162, 0.026360446587204933, 0.07530417293310165, -0.08257601410150528, -0.1432754397392273, -0.015478979796171188, 0.021100904792547226, -0.12170519679784775, -0.011143110692501068, 0.10793142765760422, -0.02897697500884533, -0.010758914984762669, 0.01802144944667816, 0.029423844069242477, -0.030877895653247833, 0.21780936419963837, 0.014865989796817303, 0.07988779246807098, -0.09235311299562454, 0.09115466475486755, 0.0558943971991539, -0.1897655427455902, 0.034690577536821365, 0.09604918211698532, -0.0694684386253357, -0.007531351409852505, 0.0656968206167221, 0.09098360687494278, 0.05718057602643967, -0.037428177893161774, -0.10310280323028564, -0.13710880279541016, 0.0753302350640297, 0.11046890914440155, 0.022716600447893143, 0.012796862982213497, -0.011631959117949009, 0.039491571485996246, -0.09593068063259125, 0.12103020399808884, 0.1091609075665474, 0.07567490637302399, -0.12966573238372803, 0.14038509130477905, 0.0013402680633589625, -0.019079651683568954, -0.012096746824681759, 0.015979932621121407, -0.11889288574457169, 0.012499586679041386, -0.10036008059978485, -0.03272233530879021, -0.061001043766736984, -0.013718636706471443, -0.010571830905973911, -0.04604179039597511, -0.03354742377996445, 0.017673460766673088, -0.10886134207248688, -0.04440770670771599, -0.021957792341709137, 0.0629410445690155, -0.09417183697223663, -0.03158898651599884, 0.0343354158103466, -0.10599327087402344, 0.08569735288619995, 0.01711456850171089, 0.03160453960299492, -0.012071683071553707, -0.10234276950359344, -0.003541304264217615, 0.028930215165019035, -0.025238189846277237, 0.03340410441160202, -0.19048474729061127, -0.018162764608860016, -0.04162639006972313, 0.007723743095993996, -0.0018079581204801798, 0.0012846989557147026, -0.11918357759714127, 0.00611121254041791, -0.04842088371515274, -0.06012127548456192, -0.052098598331213, 0.058268457651138306, 0.08396738022565842, 0.02820361591875553, 0.14741216599941254, -0.09563148766756058, 0.053415168076753616, -0.22201526165008545, -0.0006468204082921147, -0.015432536602020264, -0.08424630016088486, -0.0673205628991127, -0.027282051742076874, 0.09571380168199539, -0.05506925284862518, 0.07564152032136917, -0.04411174729466438, 0.04715763032436371, 0.04019765555858612, -0.13587111234664917, 0.022081945091485977, 0.055121053010225296, 0.1588018387556076, 0.020246226340532303, -0.02991415187716484, 0.055825892835855484, 0.011414814740419388, 0.05079081282019615, 0.14921040832996368, 0.14824171364307404, 0.15985210239887238, 0.06738953292369843, 0.06814520061016083, 0.032242462038993835, -0.10785938054323196, -0.163296177983284, 0.12382496148347855, -0.031898725777864456, 0.14727509021759033, -0.020606258884072304, 0.19107630848884583, 0.08238005638122559, -0.21827052533626556, 0.0689689964056015, -0.03925232216715813, -0.08225108683109283, -0.09262790530920029, -0.06237390264868736, -0.08273479342460632, -0.18889513611793518, 0.0037429449148476124, -0.09495288878679276, 0.05777965486049652, 0.017015870660543442, 0.03868373855948448, 0.04375274479389191, 0.12027415633201599, 0.02659645862877369, 0.011033724993467331, 0.11706036329269409, 0.02527317963540554, -0.008480163291096687, -0.027945049107074738, -0.09166756272315979, 0.05006513372063637, -0.055713169276714325, 0.04282829910516739, -0.04090418666601181, -0.09360815584659576, 0.06133235991001129, 0.03634912148118019, -0.11082321405410767, 0.023144593462347984, 0.0012720230733975768, 0.06413862854242325, 0.10881295800209045, 0.030736450105905533, 0.0019876803271472454, -0.026700420305132866, 0.23115167021751404, -0.0938059389591217, -0.03910490870475769, -0.14006899297237396, 0.23076069355010986, -0.023852698504924774, -0.006369966547936201, 0.01498311385512352, -0.07890462875366211, 0.005910450126975775, 0.13001349568367004, 0.13995316624641418, -0.013560909777879715, -0.017346113920211792, 0.014660382643342018, -0.014135737903416157, -0.033777400851249695, 0.07551665604114532, 0.09383363276720047, 0.04250920191407204, -0.05720224604010582, -0.01957225799560547, -0.03452954813838005, -0.06860727071762085, -0.02001267857849598, 0.07482288032770157, 0.03589581698179245, -0.004487794358283281, -0.020405564457178116, 0.13833768665790558, -0.029044024646282196, -0.15230053663253784, 0.046640511602163315, -0.1896083652973175, -0.17995963990688324, -0.036790259182453156, 0.07992565631866455, 0.03361279144883156, 0.048426009714603424, 0.006051590666174889, -0.013528551906347275, 0.10117912292480469, -0.005094259511679411, -0.01959017664194107, -0.11548934131860733, 0.06646192818880081, -0.1057569682598114, 0.16628246009349823, -0.041979555040597916, 0.011793295852839947, 0.11280784755945206, 0.042784903198480606, -0.08300817757844925, 0.029837852343916893, 0.07475157082080841, -0.12394076585769653, 0.036624811589717865, 0.19404259324073792, -0.0494706816971302, 0.15191563963890076, 0.048420555889606476, -0.10198178887367249, 0.03553816303610802, -0.11924412846565247, -0.06806007027626038, -0.055452611297369, 0.010180067270994186, -0.027465375140309334, 0.14371345937252045, 0.19761067628860474, -0.06031983718276024, -0.00353726907633245, -0.04735970497131348, 0.01717616245150566, 0.061451323330402374, 0.14991797506809235, -0.015732528641819954, -0.25182226300239563, 0.02709396556019783, 0.06581057608127594, 0.020631173625588417, -0.22748439013957977, -0.10352983325719833, 0.00681258412078023, -0.047539208084344864, -0.07439978420734406, 0.12388720363378525, 0.044440288096666336, 0.04413975402712822, -0.055032242089509964, -0.1244707852602005, -0.020075293257832527, 0.16574214398860931, -0.16036449372768402, -0.04567253589630127 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # bert-finetuned-ner This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.1
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "bert-base-cased", "model-index": [{"name": "bert-finetuned-ner", "results": []}]}
token-classification
yacineai7/bert-finetuned-ner
[ "transformers", "tensorboard", "safetensors", "bert", "token-classification", "generated_from_trainer", "base_model:bert-base-cased", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-10T14:35:22+00:00
[]
[]
TAGS #transformers #tensorboard #safetensors #bert #token-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
# bert-finetuned-ner This model is a fine-tuned version of bert-base-cased on an unknown dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.1
[ "# bert-finetuned-ner\n\nThis model is a fine-tuned version of bert-base-cased on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3", "### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #tensorboard #safetensors #bert #token-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "# bert-finetuned-ner\n\nThis model is a fine-tuned version of bert-base-cased on an unknown dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3", "### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1" ]
[ 68, 34, 6, 12, 8, 3, 90, 33 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #bert #token-classification #generated_from_trainer #base_model-bert-base-cased #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n# bert-finetuned-ner\n\nThis model is a fine-tuned version of bert-base-cased on an unknown dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 8\n- eval_batch_size: 8\n- seed: 42\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.0+cu121\n- Datasets 2.17.0\n- Tokenizers 0.15.1" ]
[ -0.06356771290302277, 0.08390295505523682, -0.0011577598052099347, 0.0765465721487999, 0.17440402507781982, 0.008180545642971992, 0.13111911714076996, 0.07114522904157639, -0.10896750539541245, 0.04999789223074913, 0.05961913987994194, 0.07230284810066223, 0.015765869989991188, 0.10420171171426773, -0.05041636526584625, -0.22851787507534027, 0.02167096920311451, 0.005780185107141733, -0.09927725046873093, 0.08141393214464188, 0.09285305440425873, -0.12464150786399841, 0.07085404545068741, 0.008524677716195583, -0.1884946972131729, 0.040727242827415466, 0.0024770095478743315, -0.04831158369779587, 0.10678290575742722, 0.02495172992348671, 0.14189784228801727, 0.0063184164464473724, 0.14801853895187378, -0.19814464449882507, 0.0034641542006284, 0.098280169069767, 0.03358059749007225, 0.07281583547592163, 0.05040811747312546, 0.04149753600358963, 0.06399956345558167, -0.11118552088737488, 0.09705787152051926, 0.027412299066781998, -0.05682169273495674, -0.15860509872436523, -0.0721927359700203, 0.059096578508615494, 0.10467899590730667, 0.08074402064085007, 0.007321408949792385, 0.14245407283306122, -0.08552495390176773, 0.06949537992477417, 0.1762266308069229, -0.29850712418556213, -0.08018386363983154, 0.09666257351636887, 0.0495547316968441, 0.0761103704571724, -0.10280003398656845, -0.006030090618878603, 0.063065305352211, 0.03873833268880844, 0.11987350136041641, -0.022574497386813164, -0.08895784616470337, -0.01466298196464777, -0.16562622785568237, 0.028342092409729958, 0.13095635175704956, 0.05151573568582535, -0.052304018288850784, -0.04219673573970795, -0.07369428873062134, -0.029484570026397705, -0.03914608061313629, -0.06715785712003708, 0.047546397894620895, -0.02889683097600937, -0.06006763130426407, -0.06340064853429794, -0.07479430735111237, -0.0663774162530899, -0.009889821521937847, 0.13673533499240875, 0.04792708158493042, 0.0074433814734220505, -0.028228508308529854, 0.0828462690114975, -0.037118807435035706, -0.09814880043268204, 0.029239743947982788, 0.002084678038954735, -0.045123953372240067, -0.07814773172140121, -0.05504460260272026, -0.04215576499700546, 0.025463324040174484, 0.13744591176509857, -0.048349011689424515, 0.05738183483481407, -0.005235524382442236, 0.012697900645434856, -0.05464674159884453, 0.13082201778888702, -0.05639813095331192, -0.034466587007045746, 0.03165564313530922, 0.0932575985789299, 0.03344963118433952, -0.004188512917608023, -0.11612027883529663, 0.019591227173805237, 0.09869749844074249, 0.03703044354915619, -0.0627748966217041, 0.05091336369514465, -0.014172874391078949, -0.023304112255573273, -0.000768905330915004, -0.11700241267681122, 0.04817676171660423, -0.005441570188850164, -0.05024271830916405, -0.0387008972465992, 0.020906731486320496, 0.01900884509086609, -0.006281692069023848, 0.11056485772132874, -0.09497267007827759, 0.004925283137708902, -0.11362507939338684, -0.10902400314807892, 0.009482334367930889, -0.051185447722673416, -0.0013397758593782783, -0.0922875851392746, -0.15869463980197906, -0.011315620504319668, 0.05904106795787811, -0.02669147215783596, -0.03133975714445114, -0.03479279950261116, -0.08302925527095795, 0.0009542741463519633, -0.005164760630577803, 0.12176752835512161, -0.04278651997447014, 0.05536923557519913, 0.00701348390430212, 0.0229863990098238, -0.030581455677747726, 0.030465513467788696, -0.09295322746038437, 0.030554305762052536, -0.16312438249588013, 0.024194562807679176, -0.10450585931539536, 0.03550003096461296, -0.09430532902479172, -0.08733326941728592, 0.012543758377432823, 0.0025484603829681873, 0.0644802451133728, 0.07758669555187225, -0.16465479135513306, -0.03165438771247864, 0.12796957790851593, -0.08427587151527405, -0.08280087262392044, 0.11472534388303757, -0.04393812268972397, 0.02798430435359478, 0.06034788861870766, 0.14521315693855286, 0.060850393027067184, -0.1344282627105713, -0.003735772566869855, 0.01681603491306305, 0.0779663696885109, -0.009815199300646782, 0.044428639113903046, -0.0016220619436353445, -0.028969191014766693, 0.007107457146048546, -0.0871766060590744, -0.0033043406438082457, -0.08242680132389069, -0.07744100689888, -0.052887845784425735, -0.09812991321086884, 0.0338728129863739, 0.03379080817103386, 0.043526481837034225, -0.07543899118900299, -0.11016344279050827, 0.18719038367271423, 0.10200725495815277, -0.06939049065113068, 0.014793900772929192, -0.07473919540643692, 0.060701124370098114, -0.05895305424928665, -0.018047310411930084, -0.1821439415216446, -0.13821963965892792, 0.023024966940283775, -0.04891931638121605, 0.049912046641111374, 0.05131692439317703, 0.054710812866687775, 0.09975776821374893, -0.05393295735120773, -0.0033583082258701324, -0.06780485808849335, 0.02634926326572895, -0.12257456034421921, -0.20955944061279297, -0.04954905807971954, -0.030636493116617203, 0.14006257057189941, -0.26055702567100525, 0.032342828810214996, -0.05147804319858551, 0.12337642163038254, 0.021116863936185837, -0.02410726249217987, -0.05196502059698105, 0.07450918853282928, -0.023950932547450066, -0.0807437151670456, 0.0500287264585495, 0.006111864931881428, -0.06155281141400337, -0.0926501676440239, -0.15995489060878754, 0.11252189427614212, 0.09825863689184189, -0.02712758257985115, -0.07472527772188187, 0.007073404733091593, -0.05387980863451958, -0.032148465514183044, -0.07626353949308395, 0.007710352540016174, 0.14656738936901093, -0.01317732222378254, 0.14108021557331085, -0.05977502837777138, -0.03826167434453964, 0.0005841851816512644, -0.02419048734009266, -0.004192609339952469, 0.058072615414857864, 0.12622423470020294, -0.11334707587957382, 0.11456891894340515, 0.1519192010164261, -0.11767005920410156, 0.10786525160074234, -0.03415936231613159, -0.05810638889670372, 0.0010348595678806305, -0.00785923097282648, -0.0049146185629069805, 0.12363976985216141, -0.10026606172323227, 0.0027823559939861298, 0.007254079449921846, 0.014612911269068718, 0.03528163954615593, -0.19134175777435303, 0.005992337595671415, 0.01450520008802414, -0.02421349287033081, 0.01248654443770647, -0.04657597467303276, 0.01339622214436531, 0.09321553260087967, 0.020753731951117516, -0.05790935829281807, 0.027904706075787544, 0.0023847497068345547, -0.07914050668478012, 0.19493399560451508, -0.11715411394834518, -0.1055217906832695, -0.11905384808778763, -0.01460360363125801, -0.07721390575170517, -0.005558445118367672, 0.022747932001948357, -0.09494909644126892, -0.07252644002437592, -0.09204293042421341, -0.00190201704390347, -0.007941949181258678, -0.0021465590689331293, 0.05555419251322746, 0.0003750729956664145, 0.10956686735153198, -0.1261257827281952, 0.001859403564594686, -0.03327138349413872, -0.12425442039966583, -0.018702419474720955, 0.0763213038444519, 0.10799697041511536, 0.1455809623003006, -0.033373717218637466, -0.007584606762975454, -0.020419033244252205, 0.22459232807159424, -0.03293788060545921, -0.0060217068530619144, 0.11841674149036407, -0.020599087700247765, 0.04655977711081505, 0.13828811049461365, 0.06271399557590485, -0.10546480119228363, 0.04081786051392555, 0.09655499458312988, -0.015389016829431057, -0.20614692568778992, -0.05300107225775719, -0.01961846649646759, -0.07590055465698242, 0.07722242921590805, 0.0496523380279541, 0.0065713608637452126, 0.049257438629865646, 0.006624797359108925, 0.07664166390895844, -0.013727142475545406, 0.08483721315860748, 0.1302851140499115, 0.03369176760315895, 0.12236777693033218, -0.03011099062860012, -0.045729927718639374, 0.04735884815454483, -0.009239678271114826, 0.23490941524505615, 0.010409961454570293, 0.045432597398757935, 0.06449619680643082, 0.13891847431659698, -0.013867104426026344, 0.0320427380502224, 0.0021876597311347723, -0.028230726718902588, 0.0007918324554339051, -0.06743673235177994, -0.0051653338596224785, 0.018005570396780968, -0.07752086222171783, 0.06130129471421242, -0.08651077747344971, 0.04443889483809471, 0.03941790387034416, 0.23165754973888397, 0.02085539884865284, -0.28968533873558044, -0.08304977416992188, 0.015618483535945415, -0.026308011263608932, -0.04584215208888054, 0.017571644857525826, 0.12515147030353546, -0.08464770764112473, 0.029584143310785294, -0.062131065875291824, 0.09192874282598495, 0.01084559690207243, 0.027527328580617905, 0.06732979416847229, 0.1839665025472641, -0.002270328113809228, 0.08221288025379181, -0.2328483611345291, 0.20842470228672028, 0.03072408027946949, 0.12834300100803375, -0.03778854385018349, 0.024607913568615913, 0.02756391279399395, 0.13057340681552887, 0.0650629922747612, -0.0121384859085083, -0.01923786662518978, -0.1917327642440796, -0.05044253170490265, 0.05300852656364441, 0.10962174087762833, 0.002154305810108781, 0.08437951654195786, -0.06284812837839127, 0.0006401194259524345, 0.08212079852819443, -0.04480668902397156, -0.19584190845489502, -0.10879470407962799, -0.021918511018157005, 0.01586449146270752, -0.0655071958899498, -0.09068682789802551, -0.10863162577152252, -0.09171681851148605, 0.20599690079689026, 0.004600801970809698, -0.009387579746544361, -0.1224973201751709, 0.0906558707356453, 0.06796817481517792, -0.05405784398317337, 0.03091472201049328, -0.0007255898090079427, 0.12553851306438446, 0.028204502537846565, -0.10489282011985779, 0.08189834654331207, -0.10454044491052628, -0.14747799932956696, -0.05652286857366562, 0.07782794535160065, 0.05655137449502945, 0.03910792991518974, 0.007579287048429251, 0.01505207922309637, 0.025811104103922844, -0.08570128679275513, -0.018828637897968292, 0.07881084829568863, 0.08000776916742325, 0.02332032471895218, -0.1199667751789093, -0.02343912608921528, -0.033871185034513474, 0.00034431504900567234, 0.10568556934595108, 0.20808172225952148, -0.0805840939283371, 0.020422009751200676, 0.1185271143913269, -0.08818943053483963, -0.2194192111492157, 0.09701414406299591, 0.07987026870250702, -0.006975782103836536, 0.0425214022397995, -0.17122815549373627, 0.19849549233913422, 0.12460645288228989, -0.031730249524116516, 0.049996424466371536, -0.2784743309020996, -0.13756337761878967, 0.11396749317646027, 0.15672174096107483, 0.05683767423033714, -0.14338694512844086, -0.023610036820173264, -0.04179855436086655, -0.11167386174201965, 0.13869674503803253, -0.14188556373119354, 0.0840681716799736, 0.024139899760484695, 0.0523521825671196, 0.005217447876930237, -0.04045980051159859, 0.12605004012584686, -0.005433474667370319, 0.10205338895320892, -0.05648399889469147, 0.01773231104016304, 0.08157142996788025, -0.053722675889730453, 0.01609477400779724, -0.04014371335506439, 0.05321727693080902, -0.04763927310705185, -0.027877487242221832, -0.058734215795993805, 0.08067111670970917, -0.04807126149535179, -0.07816124707460403, -0.046431370079517365, 0.04313061013817787, 0.03461937606334686, -0.031138870865106583, 0.09373395144939423, 0.024324340745806694, 0.1397407501935959, 0.10183191299438477, 0.10785714536905289, -0.07109469920396805, -0.041488658636808395, 0.013325403444468975, -0.03880101069808006, 0.08653182536363602, -0.09463873505592346, 0.029637711122632027, 0.11364483088254929, 0.023982513695955276, 0.1276966631412506, 0.06082591414451599, -0.04262028634548187, -0.011477792635560036, 0.043759942054748535, -0.12336093932390213, -0.15885217487812042, 0.0033569345250725746, -0.017062099650502205, -0.12175606191158295, 0.05900728330016136, 0.10246919095516205, -0.085057832300663, 0.0020119217224419117, -0.0077218045480549335, 0.006206423044204712, -0.04009572044014931, 0.17164863646030426, 0.05464881286025047, 0.05173100158572197, -0.07100632041692734, 0.11125253140926361, 0.06905209273099899, -0.07178512215614319, 0.036374133080244064, 0.039561204612255096, -0.09808405488729477, -0.025184115394949913, 0.09772375226020813, 0.2024209201335907, -0.04560196027159691, -0.054870251566171646, -0.09722167998552322, -0.10324768722057343, 0.02662874013185501, 0.16740281879901886, 0.06936932355165482, -0.039630793035030365, -0.04260515421628952, 0.06272929906845093, -0.14829695224761963, 0.09924538433551788, 0.007479410152882338, 0.09586263447999954, -0.17292286455631256, 0.12159088999032974, 0.023890919983386993, 0.019673854112625122, -0.0282475296407938, 0.042771510779857635, -0.12086213380098343, -0.03328714519739151, -0.16089273989200592, -0.025103101506829262, -0.02411338873207569, 0.008919570595026016, 0.0013344649923965335, -0.04326044023036957, -0.05468736216425896, 0.04990458860993385, -0.07219284027814865, -0.04243582859635353, 0.039528124034404755, 0.045017026364803314, -0.14917314052581787, 0.0038996499497443438, 0.010492702946066856, -0.07776465266942978, 0.04594266414642334, 0.022881053388118744, 0.03379547595977783, 0.06346716731786728, -0.16790536046028137, -0.02030792273581028, 0.05355233699083328, 0.04026920720934868, 0.08261995762586594, -0.05867229029536247, -0.020010491833090782, -0.004977576434612274, 0.07569822669029236, 0.004026547074317932, 0.10173514485359192, -0.10999105870723724, -0.012205722741782665, -0.06304138898849487, -0.0619894303381443, -0.045774105936288834, 0.025915058329701424, 0.10476769506931305, 0.028364859521389008, 0.19705261290073395, -0.10164979845285416, 0.005597184412181377, -0.17661774158477783, -0.024950195103883743, -0.011015670374035835, -0.0469934418797493, -0.09184417128562927, -0.037389565259218216, 0.05560650676488876, -0.06200837343931198, 0.11689355224370956, 0.01137840561568737, 0.08174505829811096, 0.04512833431363106, -0.0480852834880352, -0.00580949243158102, 0.023968789726495743, 0.21321837604045868, 0.05889420583844185, -0.007214896846562624, 0.04457271099090576, 0.009729070588946342, 0.08892641216516495, 0.037618208676576614, 0.19582054018974304, 0.15248417854309082, -0.08864115178585052, 0.0650443583726883, 0.0786748006939888, -0.09993483871221542, -0.16341489553451538, 0.09893186390399933, -0.045822035521268845, 0.12374380230903625, -0.059351831674575806, 0.1870637983083725, 0.11031018942594528, -0.15167546272277832, 0.01823471486568451, -0.06949957460165024, -0.09677136689424515, -0.13645897805690765, -0.038575541228055954, -0.08991911262273788, -0.12440390884876251, 0.009607757441699505, -0.12407855689525604, 0.008955254219472408, 0.11110763251781464, 0.007091708946973085, 0.01215338334441185, 0.1872250735759735, -0.04656430333852768, 0.04065198078751564, 0.018452012911438942, 0.00019988376880064607, -0.031103193759918213, -0.06418369710445404, -0.07705143094062805, 0.028488101437687874, 0.007896045222878456, 0.054423023015260696, -0.05751912668347359, -0.012501350603997707, 0.04730047658085823, -0.009744846262037754, -0.05973121523857117, 0.031136006116867065, 0.021642299368977547, 0.02643756754696369, 0.04561518505215645, 0.026896882802248, -0.019195258617401123, -0.02591603435575962, 0.2636190354824066, -0.08521972596645355, -0.10163721442222595, -0.11513184010982513, 0.2660549283027649, 0.015861751511693, 0.011779381893575191, 0.04091938957571983, -0.10441058874130249, -0.0026189377531409264, 0.18966786563396454, 0.15991325676441193, -0.08891458064317703, -0.014140307903289795, -0.007374660577625036, -0.025616629049181938, -0.0981825515627861, 0.15179146826267242, 0.11677615344524384, 0.035295456647872925, -0.05076124891638756, -0.0476268045604229, -0.018788781017065048, -0.016709769144654274, -0.1072002649307251, 0.032852303236722946, 0.037366487085819244, 0.005702769383788109, -0.027112707495689392, 0.05064648017287254, 0.012368503957986832, -0.1839791238307953, 0.01972014829516411, -0.11663556098937988, -0.16267341375350952, -0.03205625340342522, 0.10725368559360504, -0.045339085161685944, 0.04256926849484444, -0.03245103731751442, 0.0007198446546681225, 0.10337784886360168, -0.029007887467741966, -0.02250830829143524, -0.09565424174070358, 0.08655901998281479, -0.06696811318397522, 0.2544024884700775, -0.004091604612767696, 0.06151535362005234, 0.1176367998123169, 0.019149135798215866, -0.10441623628139496, 0.06873410940170288, 0.0557362362742424, -0.09927710890769958, 0.02394343540072441, 0.10452406853437424, -0.06075168773531914, 0.11483260244131088, 0.03053215891122818, -0.15544109046459198, 0.00026863967650569975, -0.057942409068346024, -0.06894606351852417, -0.058726370334625244, -0.0037858535069972277, -0.09341275691986084, 0.13881292939186096, 0.18797799944877625, -0.016100682318210602, 0.015729084610939026, -0.08678213506937027, 0.03932250663638115, 0.05492508038878441, 0.09372089058160782, -0.03386572748422623, -0.22984670102596283, 0.037806376814842224, -0.0070851338095963, -0.0005069029866717756, -0.2581147253513336, -0.07949208468198776, 0.03383827209472656, -0.0251038558781147, -0.08620429039001465, 0.086052805185318, 0.11345366388559341, 0.04045436903834343, -0.059322796761989594, -0.1476152241230011, -0.07413530349731445, 0.14659228920936584, -0.14248286187648773, -0.06134791672229767 ]
null
null
null
# **Q-Learning** Agent playing1 **FrozenLake-v1** This is a trained model of a **Q-Learning** agent playing **FrozenLake-v1** . ## Usage ```python model = load_from_hub(repo_id="basab1142/q-FrozenLake-v1-4x4-noSlippery", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
{"tags": ["FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "q-FrozenLake-v1-4x4-noSlippery", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "FrozenLake-v1-4x4-no_slippery", "type": "FrozenLake-v1-4x4-no_slippery"}, "metrics": [{"type": "mean_reward", "value": "1.00 +/- 0.00", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
basab1142/q-FrozenLake-v1-4x4-noSlippery
[ "FrozenLake-v1-4x4-no_slippery", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
2024-02-10T14:35:38+00:00
[]
[]
TAGS #FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
# Q-Learning Agent playing1 FrozenLake-v1 This is a trained model of a Q-Learning agent playing FrozenLake-v1 . ## Usage
[ "# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage" ]
[ "TAGS\n#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n", "# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage" ]
[ 40, 39 ]
[ "passage: TAGS\n#FrozenLake-v1-4x4-no_slippery #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 FrozenLake-v1\n This is a trained model of a Q-Learning agent playing FrozenLake-v1 .\n\n ## Usage" ]
[ 0.04578453302383423, -0.08074592798948288, -0.00430759321898222, 0.10720831900835037, 0.05034215748310089, -0.040469273924827576, 0.11997015029191971, 0.018999949097633362, 0.20601962506771088, -0.010012076236307621, 0.1455274522304535, 0.007022971753031015, -0.006192410364747047, 0.1867983490228653, 0.04572829231619835, -0.26324528455734253, 0.01831899583339691, -0.09495259821414948, -0.07281816750764847, 0.11870454251766205, 0.05470194295048714, -0.01901467889547348, -0.0007633853238075972, 0.056141503155231476, -0.0673527717590332, 0.0007737681735306978, 0.031996939331293106, -0.012976245954632759, 0.19804789125919342, -0.02254498563706875, 0.06641989201307297, 0.054705578833818436, 0.0758768692612648, -0.1998077929019928, 0.0358855277299881, -0.04215473681688309, -0.09439758956432343, -0.03934839740395546, -0.018780618906021118, 0.05878105387091637, 0.053356342017650604, 0.03858819976449013, 0.058354366570711136, 0.09384993463754654, -0.0773480236530304, 0.04328357055783272, 0.04280758649110794, 0.024811049923300743, 0.04589218273758888, -0.0237203948199749, -0.027002155780792236, 0.08246652781963348, -0.22182892262935638, 0.10318073630332947, -0.010159241035580635, -0.5270710587501526, -0.00633762264624238, 0.24088262021541595, 0.11517096310853958, 0.05707438662648201, -0.06903956830501556, 0.10566288232803345, 0.03913382440805435, -0.007209456991404295, 0.03210983797907829, 0.02150118350982666, 0.12817370891571045, 0.06009242683649063, -0.09581366181373596, 0.040699947625398636, 0.13722525537014008, 0.012822695076465607, 0.020306183025240898, -0.08888901025056839, 0.0410032719373703, -0.03461858257651329, -0.007679527159780264, -0.09758518636226654, 0.05478060990571976, 0.012466507963836193, -0.0934976264834404, -0.09247440844774246, -0.04236573353409767, -0.06708304584026337, 0.11252415925264359, 0.046419668942689896, -0.0874939113855362, 0.03884070739150047, -0.06760413944721222, 0.05918780341744423, -0.16863860189914703, 0.02074250765144825, -0.06627868115901947, -0.09376336634159088, -0.11799788475036621, -0.01683047041296959, -0.07946427166461945, 0.009092256426811218, 0.056664444506168365, 0.1447116881608963, 0.22076484560966492, 0.06690320372581482, 0.09728849679231644, 0.07456006109714508, 0.06531001627445221, 0.1538129299879074, 0.10918238013982773, 0.019075315445661545, -0.015266558155417442, 0.0948706716299057, -0.06445580720901489, -0.1351388692855835, -0.15579092502593994, 0.005488025024533272, 0.0983937531709671, 0.08871900290250778, -0.044080477207899094, -0.006702381651848555, -0.024641724303364754, 0.08566431701183319, -0.11314457654953003, -0.024612564593553543, -0.002267979085445404, 0.06882024556398392, -0.024801667779684067, 0.020378148183226585, -0.06242705136537552, 0.12715265154838562, 0.04222423583269119, -0.059924717992544174, -0.055308472365140915, -0.03053177334368229, -0.014276440255343914, -0.027539284899830818, 0.02446848154067993, -0.07659092545509338, 0.04767750948667526, -0.16766095161437988, -0.042871296405792236, -0.04784649610519409, 0.025697942823171616, -0.03907240927219391, -0.13557587563991547, -0.17699143290519714, -0.048906855285167694, -0.022438718006014824, 0.03549358621239662, -0.038111843168735504, 0.006551501806825399, -0.006318534724414349, -0.1583600640296936, 0.09783563017845154, 0.09784027189016342, -0.03643378987908363, -0.02749447710812092, 0.056263517588377, -0.07194498926401138, 0.1561182290315628, -0.21054518222808838, -0.054014235734939575, -0.044764336198568344, -0.06595750898122787, 0.19673264026641846, 0.012690845876932144, -0.01202624011784792, 0.19873127341270447, -0.29073721170425415, -0.06078760325908661, 0.12533614039421082, -0.07834373414516449, -0.0936407670378685, 0.06941844522953033, -0.04206686094403267, 0.023345354944467545, 0.046047765761613846, 0.36345911026000977, -0.02069227211177349, -0.16197136044502258, -0.021782705560326576, 0.13971707224845886, -0.1184760183095932, 0.059895481914281845, 0.04240793362259865, 0.12543781101703644, -0.04250509291887283, -0.018672896549105644, -0.09023164212703705, 0.05999075248837471, -0.05241934582591057, -0.09016361832618713, -0.03393383324146271, -0.07645075023174286, 0.13294468820095062, -0.0629684180021286, 0.05601520463824272, -0.03255095332860947, -0.07133250683546066, -0.050324998795986176, -0.016492370516061783, 0.04460815340280533, 0.05951254442334175, -0.12794871628284454, 0.11029167473316193, 0.13025271892547607, -0.0006193425506353378, -0.07498852163553238, -0.17872096598148346, 0.003240168560296297, 0.009576505981385708, 0.039837226271629333, 0.17141658067703247, 0.12209978699684143, 0.033295199275016785, 0.008770671673119068, -0.06389404833316803, -0.18276847898960114, 0.058129217475652695, -0.056212130934000015, -0.14230976998806, -0.052409034222364426, -0.0728459507226944, 0.017381802201271057, -0.0859743058681488, -0.017379917204380035, 0.021926190704107285, 0.006908397190272808, 0.02990424446761608, -0.026645656675100327, -0.049561817198991776, 0.021254703402519226, 0.06490101665258408, -0.0037617047782987356, 0.12023693323135376, 0.008277264423668385, -0.18308481574058533, 0.07930773496627808, 0.08478537946939468, 0.09196605533361435, 0.013250201940536499, 0.02685922384262085, -0.021522263064980507, -0.08061408251523972, -0.054420311003923416, 0.02957955375313759, 0.11417073011398315, 0.1317172348499298, 0.2361993044614792, 0.08753683418035507, 0.04697408527135849, -0.02164587564766407, -0.016415923833847046, 0.002810494042932987, -0.06318057328462601, -0.029935607686638832, 0.10614971816539764, 0.05865858122706413, -0.067733034491539, -0.04576427489519119, 0.09590928256511688, 0.02732124738395214, 0.21205885708332062, -0.03342745825648308, 0.01286078616976738, -0.10957037657499313, -0.06550975888967514, -0.031982194632291794, 0.09201868623495102, 0.09498392790555954, 0.009755023755133152, -0.022056059911847115, -0.04259001836180687, 0.0012916827108711004, -0.1334889680147171, -0.10375088453292847, 0.026475343853235245, 0.013400445692241192, -0.11206940561532974, 0.11674030870199203, -0.11352457851171494, 0.039504457265138626, 0.06024791672825813, -0.13837239146232605, 0.04428480193018913, -0.029713207855820656, -0.07886212319135666, 0.16866780817508698, -0.11075661331415176, -0.094340018928051, -0.08831550180912018, 0.004082420375198126, 0.0075836325995624065, -0.03922267258167267, -0.009283260442316532, -0.19952571392059326, -0.005375816952437162, -0.03544965013861656, 0.013616434298455715, -0.06988783925771713, -0.11287739872932434, -0.010957922786474228, 0.07084179669618607, -0.043388739228248596, -0.07803605496883392, 0.007967432029545307, -0.08923084288835526, -0.10623309016227722, 0.028189711272716522, 0.019765101373195648, -0.022883659228682518, 0.16152891516685486, 0.01816628873348236, 0.05626589432358742, -0.03298520669341087, 0.30665266513824463, -0.038163769990205765, 0.08371731638908386, -0.02993497997522354, -0.07433546334505081, 0.06130730360746384, -0.022327827289700508, 0.06086638569831848, -0.020221687853336334, -0.02362890914082527, 0.0077952733263373375, -0.08579335361719131, -0.18365982174873352, -0.05417544022202492, 0.03724347800016403, 0.195254847407341, 0.031118987128138542, 0.01910330168902874, -0.0488768145442009, -0.010547760874032974, 0.1665220558643341, -0.10005921125411987, 0.04030545800924301, -0.05366240441799164, 0.11506262421607971, -0.08640182018280029, 0.06195629760622978, 0.020486772060394287, 0.04266135022044182, -0.04877188801765442, 0.09486009180545807, 0.0826394334435463, 0.1121082529425621, -0.02206910029053688, 0.046257395297288895, 0.019012698903679848, 0.07383184134960175, 0.11073657125234604, 0.0368414968252182, -0.0729052945971489, 0.001982470043003559, -0.006313489284366369, -0.039427030831575394, 0.11933320760726929, 0.17963355779647827, -0.11991413682699203, -0.05106910318136215, 0.27167606353759766, 0.0031242913100868464, 0.19481229782104492, -0.01315275114029646, 0.043591804802417755, -0.04484925419092178, 0.04572054371237755, -0.05338600277900696, -0.04086209088563919, 0.2094656229019165, 0.08045925945043564, -0.17165091633796692, -0.08549032360315323, -0.05912299454212189, 0.07081323862075806, 0.10728751868009567, 0.0013539529172703624, -0.04156802222132683, 0.0004610282776411623, 0.0014198932331055403, 0.08339415490627289, -0.14520122110843658, 0.11816094070672989, -0.03172019124031067, 0.05612684786319733, 0.017555562779307365, -0.045326150953769684, 0.04264266416430473, 0.07474290579557419, 0.26618310809135437, 0.0904107540845871, -0.040318213403224945, -0.0892091691493988, -0.12260187417268753, 0.010461576282978058, 0.029102616012096405, -0.03534553572535515, 0.0037547778338193893, -0.020087555050849915, 0.0318896509706974, 0.008264793083071709, 0.016230624169111252, -0.08987458795309067, -0.03175399824976921, -0.027736429125070572, -0.023839212954044342, 0.10733365267515182, -0.09495144337415695, -0.1444292515516281, -0.15713949501514435, 0.04191131144762039, -0.0766405463218689, -0.056593164801597595, -0.054507751017808914, -0.05239389091730118, -0.0311186034232378, -0.03773957118391991, 0.09099467098712921, -0.0021037792321294546, 0.14807306230068207, -0.1920108050107956, -0.04220759496092796, 0.051812779158353806, -0.07607918977737427, -0.08729588985443115, 0.03410962224006653, 0.12136995792388916, 0.05116051807999611, 0.11504370719194412, 0.013609255664050579, 0.09567681699991226, 0.0045484392903745174, -0.06713183224201202, 0.15302421152591705, -0.14069625735282898, -0.27875974774360657, -0.03836318850517273, 0.016946332529187202, 0.1615200787782669, -0.05613167956471443, 0.031766023486852646, 0.3335736393928528, 0.27782970666885376, -0.1428707242012024, 0.25916144251823425, 0.019178593531250954, 0.004398873541504145, -0.19130495190620422, -0.10125631093978882, 0.025324683636426926, 0.04740457236766815, 0.12032642960548401, -0.14564448595046997, -0.010732659138739109, -0.04543145373463631, -0.025908485054969788, 0.10386138409376144, -0.12300799041986465, -0.07263197749853134, 0.07765276730060577, 0.039809420704841614, 0.1808302253484726, 0.03932500258088112, 0.0014799144119024277, 0.13626977801322937, 0.06612244248390198, 0.019124457612633705, 0.05216038227081299, 0.08028066903352737, -0.018944554030895233, 0.14207926392555237, 0.05448179319500923, -0.02551644667983055, 0.052681710571050644, -0.0054580713622272015, -0.03219012916088104, 0.015605825930833817, -0.183198019862175, -0.10147556662559509, -0.0561356320977211, -0.10798973590135574, -0.04978342354297638, 0.056853994727134705, -0.12395523488521576, -0.007896827533841133, -0.03841273859143257, 0.03718273714184761, -0.07831971347332001, -0.09360362589359283, -0.036494381725788116, 0.1351792961359024, 0.07210618257522583, 0.04471297934651375, 0.035655103623867035, -0.07390819489955902, 0.07097936421632767, 0.21671734750270844, 0.08159157633781433, 0.028919655829668045, -0.19545674324035645, -0.024042490869760513, -0.0803457647562027, 0.06306298077106476, -0.08856996893882751, -0.016788700595498085, 0.11923003196716309, 0.08616556972265244, 0.05413002520799637, 0.09640096127986908, -0.045083072036504745, 0.021686913445591927, 0.02684609219431877, -0.15131035447120667, -0.18501274287700653, -0.08534606546163559, -0.03519878163933754, 0.11561143398284912, -0.06398691236972809, 0.10897188633680344, -0.13615410029888153, 0.010051886551082134, -0.006060056854039431, 0.02693452313542366, -0.03596206381917, -0.11251141875982285, 0.15348562598228455, 0.11999429017305374, -0.06767056882381439, 0.03127254918217659, -0.09527092427015305, -0.04423454403877258, 0.12686803936958313, -0.013623855076730251, -0.0371493324637413, -0.054547641426324844, -0.03628576174378395, 0.15247689187526703, -0.03436964750289917, 0.008244883269071579, -0.041229065507650375, -0.18217355012893677, 0.0798322781920433, 0.09045056998729706, 0.019827889278531075, -0.031874191015958786, -0.09797266125679016, -0.010231015272438526, -0.0011165260802954435, 0.11730700731277466, -0.10696814209222794, -0.10933240503072739, -0.15144047141075134, 0.06713984161615372, -0.0007159380475059152, 0.18502596020698547, -0.06394898891448975, -0.08904669433832169, -0.12429379671812057, 0.02344517596065998, -0.0027384376153349876, -0.042264558374881744, 0.01618490368127823, 0.07992301136255264, -0.04095321521162987, 0.02075677551329136, -0.06651144474744797, 0.06372585147619247, -0.11786920577287674, 0.09625071287155151, 0.01063506118953228, 0.016993753612041473, -0.0417880080640316, -0.01618220843374729, 0.039470795542001724, -0.057925306260585785, 0.07921463251113892, 0.011758086271584034, 0.0010938759660348296, 0.10196787863969803, -0.0034960443153977394, 0.06409632414579391, -0.05372481048107147, -0.023290161043405533, 0.06578411161899567, -0.05874887853860855, -0.03370826691389084, -0.1573946475982666, -0.0709633082151413, 0.020051732659339905, -0.04775108024477959, 0.002077929675579071, 0.03673801198601723, 0.062159497290849686, -0.06937079131603241, -0.12125655263662338, -0.043812792748212814, -0.028638383373618126, 0.021301284432411194, 0.10829301923513412, -0.07526551932096481, 0.1547859013080597, -0.052787959575653076, -0.00020603960729204118, 0.07437096536159515, 0.04048224538564682, 0.01393822580575943, -0.10422444343566895, -0.04698587954044342, -0.11035211384296417, 0.1502903699874878, -0.007902312092483044, -0.03533121198415756, 0.03719403222203255, -0.11946307867765427, -0.1572723090648651, 0.03418220207095146, 0.10199101269245148, 0.0448341928422451, 0.025807438418269157, 0.027079269289970398, -0.04042419046163559, -0.021270349621772766, -0.07034418731927872, 0.0882953479886055, -0.12085357308387756, -0.09669415652751923, 0.09555385261774063, 0.12178351730108261, -0.0036850625183433294, -0.07441367954015732, 0.11554073542356491, -0.021787192672491074, 0.05525410920381546, -0.02971339225769043, 0.10308072715997696, 0.0796005055308342, -0.12273547053337097, 0.005693064536899328, -0.036891788244247437, -0.0741485133767128, -0.12975730001926422, 0.019545545801520348, -0.061916105449199677, -0.13383042812347412, 0.12179028987884521, -0.09376577287912369, 0.030037038028240204, -0.10506992787122726, 0.021338803693652153, 0.01864001713693142, 0.061665527522563934, -0.10988292098045349, 0.08575301617383957, 0.13424484431743622, -0.043199893087148666, -0.07184189558029175, -0.12455986440181732, -0.05022053420543671, -0.04231856390833855, -0.13957437872886658, -0.11600435525178909, 0.0100301094353199, -0.023418782278895378, -0.05818291753530502, 0.0015462689334526658, -0.03659068048000336, 0.008594646118581295, 0.021907730028033257, 0.04032021388411522, -0.02693161368370056, 0.05134565755724907, -0.057569269090890884, -0.052510857582092285, 0.11489357799291611, 0.04113486409187317, -0.03561042994260788, -0.052359987050294876, 0.12997733056545258, -0.11959461867809296, 0.07662346214056015, -0.020313527435064316, 0.017129231244325638, -0.06435854732990265, 0.17131924629211426, 0.11673715710639954, -0.1367570012807846, -0.005008010193705559, -0.08210669457912445, 0.020409544929862022, 0.023555370047688484, 0.13693512976169586, -0.03411718085408211, -0.0012358218664303422, -0.1580323874950409, 0.018575575202703476, -0.18557456135749817, -0.03716109320521355, 0.04671547934412956, 0.09917585551738739, 0.15293832123279572, -0.0034432117827236652, -0.1263325810432434, 0.10424192249774933, -0.2118520885705948, 0.0907607227563858, 0.05121984705328941, -0.11874113976955414, -0.06765396893024445, -0.06795281916856766, 0.1198519766330719, 0.009196433238685131, 0.2040700763463974, -0.013615905307233334, -0.09132910519838333, -0.07060808688402176, -0.01980910450220108, -0.030524181202054024, 0.09714830666780472, 0.041414931416511536, 0.04653804749250412, 0.12821412086486816, 0.00368314771912992, 0.07533777505159378, 0.060310911387205124, 0.02759413793683052, -0.012300663627684116, 0.04076618701219559, 0.08261215686798096, -0.14588621258735657, -0.1659701019525528, 0.1326720416545868, 0.025149408727884293, 0.11792458593845367, 0.03658788278698921, -0.1549617499113083, 0.06687124073505402, 0.2523096203804016, -0.11147607117891312, 0.02505038119852543, 0.12737524509429932, -0.0366884209215641, 0.0672016367316246, 0.1144871786236763, -0.02633814327418804, -0.05217865854501724, -0.011363590136170387, 0.10233135521411896, 0.028660254552960396, -0.04646271467208862, -0.02340836264193058, -0.03373933956027031, -0.019070526584982872, -0.011738128960132599, -0.0909019410610199, -0.1543993502855301, -0.10471053421497345, -0.16619662940502167, 0.04399140924215317, -0.04626438021659851, 0.13418889045715332, 0.09469578415155411, -0.012723101302981377, 0.04568437114357948, 0.028575526550412178, 0.07275456190109253, 0.07916246354579926, -0.02939477376639843, -0.036159269511699677 ]
null
null
null
<br> # Vim Model Card ## Model Details Vision Mamba (Vim) is a generic backbone trained on the ImageNet-1K dataset for vision tasks. - **Developed by:** [HUST](https://english.hust.edu.cn/), [Horizon Robotics](https://en.horizon.cc/), [BAAI](https://www.baai.ac.cn/english.html) - **Model type:** A generic vision backbone based on the bidirectional state space model (SSM) architecture. - **License:** Non-commercial license ### Model Sources - **Repository:** https://github.com/hustvl/Vim - **Paper:** https://arxiv.org/abs/2401.09417 ## Uses The primary use of Vim is research on vision tasks, e.g., classification, segmentation, detection, and instance segmentation, with an SSM-based backbone. The primary intended users of the model are researchers and hobbyists in computer vision, machine learning, and artificial intelligence. ## How to Get Started with the Model - You can replace the backbone for vision tasks with the proposed Vim: https://github.com/hustvl/Vim/blob/main/vim/models_mamba.py - Then you can load this checkpoint and start training. ## Training Details Vim is pretrained on ImageNet-1K with classification supervision. The training data is around 1.3M images from [ImageNet-1K dataset](https://www.image-net.org/challenges/LSVRC/2012/). See more details in this [paper](https://arxiv.org/abs/2401.09417). ## Evaluation Vim-small is evaluated on ImageNet-1K val set, and achieves 80.5% Top-1 Acc. By further finetuning at finer granularity, Vim-small achieves 81.6% Top-1 Acc. See more details in this [paper](https://arxiv.org/abs/2401.09417). ## Additional Information ### Citation Information ``` @article{vim, title={Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model}, author={Lianghui Zhu and Bencheng Liao and Qian Zhang and Xinlong Wang and Wenyu Liu and Xinggang Wang}, journal={arXiv preprint arXiv:2401.09417}, year={2024} } ```
{"license": "apache-2.0"}
null
hustvl/Vim-small-midclstok
[ "arxiv:2401.09417", "license:apache-2.0", "region:us" ]
2024-02-10T14:36:20+00:00
[ "2401.09417" ]
[]
TAGS #arxiv-2401.09417 #license-apache-2.0 #region-us
<br> # Vim Model Card ## Model Details Vision Mamba (Vim) is a generic backbone trained on the ImageNet-1K dataset for vision tasks. - Developed by: HUST, Horizon Robotics, BAAI - Model type: A generic vision backbone based on the bidirectional state space model (SSM) architecture. - License: Non-commercial license ### Model Sources - Repository: URL - Paper: URL ## Uses The primary use of Vim is research on vision tasks, e.g., classification, segmentation, detection, and instance segmentation, with an SSM-based backbone. The primary intended users of the model are researchers and hobbyists in computer vision, machine learning, and artificial intelligence. ## How to Get Started with the Model - You can replace the backbone for vision tasks with the proposed Vim: URL - Then you can load this checkpoint and start training. ## Training Details Vim is pretrained on ImageNet-1K with classification supervision. The training data is around 1.3M images from ImageNet-1K dataset. See more details in this paper. ## Evaluation Vim-small is evaluated on ImageNet-1K val set, and achieves 80.5% Top-1 Acc. By further finetuning at finer granularity, Vim-small achieves 81.6% Top-1 Acc. See more details in this paper. ## Additional Information
[ "# Vim Model Card", "## Model Details\n\nVision Mamba (Vim) is a generic backbone trained on the ImageNet-1K dataset for vision tasks.\n\n- Developed by: HUST, Horizon Robotics, BAAI\n- Model type: A generic vision backbone based on the bidirectional state space model (SSM) architecture.\n- License: Non-commercial license", "### Model Sources\n\n- Repository: URL\n- Paper: URL", "## Uses\n\nThe primary use of Vim is research on vision tasks, e.g., classification, segmentation, detection, and instance segmentation, with an SSM-based backbone.\nThe primary intended users of the model are researchers and hobbyists in computer vision, machine learning, and artificial intelligence.", "## How to Get Started with the Model\n\n- You can replace the backbone for vision tasks with the proposed Vim: URL\n- Then you can load this checkpoint and start training.", "## Training Details\n\nVim is pretrained on ImageNet-1K with classification supervision.\nThe training data is around 1.3M images from ImageNet-1K dataset.\nSee more details in this paper.", "## Evaluation\n\nVim-small is evaluated on ImageNet-1K val set, and achieves 80.5% Top-1 Acc. By further finetuning at finer granularity, Vim-small achieves 81.6% Top-1 Acc. See more details in this paper.", "## Additional Information" ]
[ "TAGS\n#arxiv-2401.09417 #license-apache-2.0 #region-us \n", "# Vim Model Card", "## Model Details\n\nVision Mamba (Vim) is a generic backbone trained on the ImageNet-1K dataset for vision tasks.\n\n- Developed by: HUST, Horizon Robotics, BAAI\n- Model type: A generic vision backbone based on the bidirectional state space model (SSM) architecture.\n- License: Non-commercial license", "### Model Sources\n\n- Repository: URL\n- Paper: URL", "## Uses\n\nThe primary use of Vim is research on vision tasks, e.g., classification, segmentation, detection, and instance segmentation, with an SSM-based backbone.\nThe primary intended users of the model are researchers and hobbyists in computer vision, machine learning, and artificial intelligence.", "## How to Get Started with the Model\n\n- You can replace the backbone for vision tasks with the proposed Vim: URL\n- Then you can load this checkpoint and start training.", "## Training Details\n\nVim is pretrained on ImageNet-1K with classification supervision.\nThe training data is around 1.3M images from ImageNet-1K dataset.\nSee more details in this paper.", "## Evaluation\n\nVim-small is evaluated on ImageNet-1K val set, and achieves 80.5% Top-1 Acc. By further finetuning at finer granularity, Vim-small achieves 81.6% Top-1 Acc. See more details in this paper.", "## Additional Information" ]
[ 23, 5, 78, 15, 69, 40, 43, 62, 5 ]
[ "passage: TAGS\n#arxiv-2401.09417 #license-apache-2.0 #region-us \n# Vim Model Card## Model Details\n\nVision Mamba (Vim) is a generic backbone trained on the ImageNet-1K dataset for vision tasks.\n\n- Developed by: HUST, Horizon Robotics, BAAI\n- Model type: A generic vision backbone based on the bidirectional state space model (SSM) architecture.\n- License: Non-commercial license### Model Sources\n\n- Repository: URL\n- Paper: URL## Uses\n\nThe primary use of Vim is research on vision tasks, e.g., classification, segmentation, detection, and instance segmentation, with an SSM-based backbone.\nThe primary intended users of the model are researchers and hobbyists in computer vision, machine learning, and artificial intelligence.## How to Get Started with the Model\n\n- You can replace the backbone for vision tasks with the proposed Vim: URL\n- Then you can load this checkpoint and start training.## Training Details\n\nVim is pretrained on ImageNet-1K with classification supervision.\nThe training data is around 1.3M images from ImageNet-1K dataset.\nSee more details in this paper.## Evaluation\n\nVim-small is evaluated on ImageNet-1K val set, and achieves 80.5% Top-1 Acc. By further finetuning at finer granularity, Vim-small achieves 81.6% Top-1 Acc. See more details in this paper.## Additional Information" ]
[ -0.07090766727924347, 0.09033099561929703, -0.004465581849217415, 0.004966770764440298, 0.10628427565097809, 0.031193861737847328, 0.079409159719944, 0.08659134060144424, -0.023941799998283386, 0.028028422966599464, 0.04957452788949013, 0.024355659261345863, 0.10880967974662781, 0.1787625402212143, 0.08979351073503494, -0.19492805004119873, 0.009274208918213844, -0.03797217085957527, 0.03698456287384033, 0.04573714733123779, 0.12114256620407104, -0.09212882816791534, 0.04919460788369179, -0.009086066856980324, -0.12221816927194595, 0.02644210495054722, 0.0007874759030528367, 0.015253742225468159, 0.13671794533729553, -0.004851363133639097, 0.07688675075769424, -0.03616596758365631, 0.10078778862953186, -0.13560746610164642, 0.027216115966439247, 0.09440263360738754, 0.04614555090665817, 0.06188849359750748, 0.056635089218616486, 0.1331559270620346, 0.14252248406410217, -0.05857197195291519, 0.02393682673573494, 0.03640054911375046, -0.0186675563454628, -0.1097119152545929, -0.11694823205471039, 0.07699763774871826, 0.10574612766504288, 0.05320984497666359, -0.007834176532924175, 0.11205644905567169, 0.04810969531536102, 0.06055484339594841, -0.0005354689783416688, -0.26052621006965637, -0.06057913973927498, 0.18718630075454712, -0.0062538995407521725, 0.05177019163966179, -0.05146194249391556, 0.04825699329376221, -0.006537287961691618, 0.022489851340651512, 0.05576936528086662, 0.03955692797899246, -0.04311579838395119, -0.039889607578516006, -0.14102457463741302, -0.04480110481381416, 0.13685691356658936, 0.061575230211019516, -0.04952133819460869, -0.09851889312267303, -0.07327628135681152, 0.03697739541530609, 0.018145965412259102, -0.03686254099011421, 0.028224967420101166, 0.024889487773180008, 0.13100455701351166, -0.08000440895557404, -0.11772759258747101, -0.02806638926267624, 0.027002930641174316, -0.06810442358255386, 0.044595103710889816, 0.06306198239326477, 0.03803881257772446, 0.09240738302469254, -0.11158648878335953, -0.07182499021291733, -0.020991191267967224, -0.05792431905865669, -0.08008391410112381, -0.05842912942171097, 0.02949601039290428, -0.22131679952144623, -0.03877351060509682, 0.038021814078092575, 0.07124543935060501, 0.07054658234119415, -0.03305564820766449, 0.0428660549223423, 0.0330246239900589, 0.10405184328556061, -0.10806091874837875, 0.006081092171370983, 0.06461833417415619, 0.023539196699857712, 0.03390355408191681, -0.000697238021530211, -0.021899523213505745, 0.0034311956260353327, 0.01940661296248436, 0.019518442451953888, -0.018043305724859238, 0.0027072320226579905, 0.030002249404788017, -0.04215892776846886, 0.11165574193000793, -0.1110694482922554, 0.016138620674610138, 0.0001277098635910079, -0.03175957500934601, 0.020592181012034416, 0.0545436292886734, -0.01782829314470291, -0.10190724581480026, -0.01619890332221985, -0.011766614392399788, -0.03924837335944176, -0.10406658053398132, -0.08754605054855347, 0.05939215421676636, -0.1454319804906845, -0.06065835431218147, -0.10835321992635727, -0.08400054275989532, -0.054956283420324326, 0.13985474407672882, -0.07446569949388504, -0.021814128383994102, -0.02848149463534355, -0.043020252138376236, 0.008791734464466572, 0.05216600000858307, 0.131522998213768, 0.012794272974133492, 0.0029735760763287544, -0.04656965658068657, 0.08528254181146622, 0.054847601801157, -0.0016210228204727173, -0.005813851952552795, 0.007475252263247967, -0.1189948171377182, 0.011955220252275467, -0.048491109162569046, 0.04964449629187584, -0.17173494398593903, 0.023380020633339882, -0.04462674260139465, -0.011875673197209835, 0.08782285451889038, 0.0921880304813385, -0.20891346037387848, 0.04214054346084595, 0.04947901889681816, -0.12448428571224213, -0.08156999200582504, 0.035305771976709366, 0.005554496310651302, 0.12333298474550247, 0.023240813985466957, 0.09593891352415085, 0.06107456609606743, -0.15760111808776855, -0.023536909371614456, 0.009093822911381721, -0.05089873448014259, -0.07729947566986084, 0.04404372349381447, 0.09725112468004227, -0.04205082729458809, 0.010822173207998276, -0.12921078503131866, 0.02138173021376133, -0.022199120372533798, -0.08326498419046402, -0.04332106187939644, -0.0376252681016922, -0.09738500416278839, 0.021949557587504387, 0.032919108867645264, 0.039245910942554474, -0.06179952621459961, -0.04437556490302086, 0.08919887244701385, -0.015025242231786251, 0.010882929898798466, -0.09832853078842163, 0.017022307962179184, -0.14506570994853973, 0.05805148184299469, -0.13268400728702545, 0.03347908705472946, 0.058425821363925934, -0.034381281584501266, 0.08248303830623627, 0.05142026022076607, 0.037705183029174805, 0.007801560219377279, -0.037404172122478485, -0.046491269022226334, -0.0021587456576526165, -0.026985228061676025, -0.030973317101597786, -0.20149436593055725, -0.06358372420072556, -0.04971427097916603, -0.041002750396728516, -0.16880525648593903, -0.0019747342448681593, 0.011472207494080067, 0.1302933394908905, 0.03315715864300728, -0.05989169701933861, 0.02099424973130226, -0.032681819051504135, -0.008443033322691917, -0.05031443014740944, 0.05048077553510666, 0.02906705252826214, 0.04115648940205574, 0.012660427950322628, 0.024510880932211876, -0.11746276915073395, 0.058456480503082275, -0.037285320460796356, -0.0597531758248806, 0.07518228143453598, -0.011292610317468643, -0.018576793372631073, -0.056932203471660614, 0.0311423409730196, 0.017749691382050514, 0.05331537127494812, 0.07693161070346832, -0.04925662279129028, -0.02602638490498066, 0.046586472541093826, 0.009937865659594536, -0.07805187255144119, 0.025512952357530594, 0.2247973382472992, -0.19997744262218475, -0.027179433032870293, 0.018039630725979805, 0.0011397019261494279, 0.07093504071235657, 0.009848934598267078, -0.05167229846119881, -0.04619275778532028, 0.03133026883006096, 0.013567754067480564, 0.16246655583381653, -0.13506828248500824, -0.0175126101821661, 0.02163468301296234, 0.013535311445593834, 0.012368550524115562, -0.12921877205371857, 0.030369630083441734, 0.008157903328537941, 0.001974395476281643, 0.02501233108341694, 0.011766927316784859, -0.07749780267477036, 0.06183413416147232, 0.08785803616046906, 0.0476297028362751, -0.05535870045423508, -0.06758240610361099, -0.1280919909477234, 0.1229071319103241, -0.08941762149333954, -0.2354924976825714, -0.1528582125902176, 0.020963609218597412, -0.027769317850470543, -0.012343913316726685, -0.0348224975168705, -0.11423127353191376, -0.09851899743080139, -0.10814610123634338, 0.0030551832169294357, 0.009362895041704178, 0.0034407214261591434, 0.04084758833050728, 0.05438760295510292, 0.06115607172250748, -0.10971417278051376, 0.04121079295873642, -0.0030712431762367487, -0.01526786107569933, 0.040840111672878265, 0.055387500673532486, 0.027324752882122993, 0.14825299382209778, -0.03682698309421539, 0.05993981286883354, 0.00581808015704155, 0.24141111969947815, -0.05902709439396858, 0.1361294984817505, 0.1848258227109909, 0.02282940223813057, 0.00812109000980854, 0.1959989070892334, 0.025532551109790802, -0.06467166543006897, 0.025646070018410683, 0.015272043645381927, -0.03233950957655907, -0.21437960863113403, -0.002566342009231448, -0.04542519524693489, -0.09221051633358002, 0.0845225602388382, 0.09381360560655594, -0.049438491463661194, 0.03679649904370308, 0.0008864792180247605, -0.01079891249537468, 0.02798132598400116, 0.08525528013706207, 0.1202535554766655, -0.04629846662282944, 0.02464822307229042, -0.10827700048685074, 0.0826251208782196, 0.03838132694363594, 0.1180708035826683, 0.22487321496009827, 0.015677358955144882, 0.013796940445899963, 0.09822838753461838, 0.06717059761285782, -0.02064756490290165, 0.051242582499980927, -0.1047615110874176, -0.0002656459400895983, -0.011034848168492317, -0.07617983967065811, 0.014248241670429707, 0.06747855246067047, -0.1281144618988037, -0.033052727580070496, 0.0027351921889930964, 0.057480838149785995, 0.016699517145752907, 0.18737830221652985, 0.08056221902370453, -0.15322361886501312, -0.032430499792099, 0.012551534920930862, -0.07183004170656204, -0.1137462928891182, -0.051651597023010254, 0.10647092014551163, -0.02713748812675476, 0.0451638363301754, -0.04228610917925835, 0.07699917256832123, -0.0491967499256134, -0.04082852602005005, 0.07081209123134613, 0.04330037906765938, 0.04652661830186844, 0.06633693724870682, -0.03533153235912323, 0.138586163520813, 0.006647149566560984, 0.15207141637802124, -0.03485003486275673, 0.023375699296593666, -0.03624899312853813, 0.12963944673538208, 0.12063926458358765, 0.029143227264285088, -0.031845178455114365, -0.008704404346644878, 0.007518745958805084, 0.01700092852115631, 0.044610898941755295, 0.006549831014126539, 0.07545680552721024, -0.03231934830546379, -0.03402603790163994, 0.004275839310139418, -0.040890883654356, -0.16952066123485565, -0.10328320413827896, 0.019919205456972122, -0.014674864709377289, -0.022409409284591675, -0.09162207692861557, -0.05359846353530884, 0.0010791118256747723, 0.20093700289726257, -0.1410677433013916, -0.003035749541595578, -0.09516801685094833, 0.05405239760875702, 0.04213440790772438, -0.027747804298996925, 0.08003047108650208, -0.05595690757036209, 0.18382346630096436, -0.020123176276683807, -0.09164543449878693, 0.011695818044245243, -0.061659738421440125, -0.06335103511810303, -0.04593692347407341, 0.07909241318702698, -0.00887118186801672, 0.004923352040350437, 0.017677107825875282, 0.03323306143283844, 0.025453263893723488, -0.02973339334130287, 0.049634918570518494, 0.1773061901330948, -0.06973542273044586, -0.07794621586799622, -0.05723631754517555, -0.06491714715957642, -0.02735891006886959, -0.10060498118400574, 0.06494726985692978, 0.1863340139389038, -0.06194404885172844, 0.1168799102306366, 0.10185789316892624, -0.12946586310863495, -0.26423558592796326, 0.06573076546192169, 0.052512723952531815, 0.06455420702695847, -0.047497887164354324, -0.2171710878610611, 0.08107303082942963, -0.09337272495031357, -0.03523270785808563, 0.0030009883921593428, -0.22914685308933258, -0.10255041718482971, -0.039084069430828094, 0.14750847220420837, 0.05092800036072731, 0.005833595059812069, -0.014609063975512981, -0.001273648114874959, -0.010502583347260952, 0.05503842607140541, 0.08200535923242569, 0.08824051916599274, 0.0016094143502414227, -0.023552561178803444, 0.00909376796334982, -0.05363078415393829, 0.06257441639900208, -0.07962418347597122, 0.035893529653549194, -0.0223829485476017, 0.02599460445344448, 0.07657548785209656, -0.07269205898046494, 0.1486269235610962, 0.09242847561836243, 0.02765592746436596, -0.11492674052715302, -0.04793856665492058, -0.04736166447401047, 0.008937287144362926, -0.03467177227139473, -0.02535199746489525, -0.0420486144721508, 0.09493646770715714, 0.045588988810777664, -0.02819696068763733, 0.060198601335287094, -0.0017281617037951946, -0.07401606440544128, 0.14223580062389374, 0.07683251798152924, -0.0971054658293724, 0.02845585159957409, 0.061103906482458115, -0.01395496353507042, 0.09004917740821838, -0.23273974657058716, 0.006387567613273859, 0.08583276718854904, 0.0071672783233225346, 0.04073648154735565, -0.007735732011497021, -0.1333015412092209, -0.06884413212537766, 0.09876374900341034, -0.045924898236989975, -0.18877564370632172, -0.004207275342196226, 0.15340083837509155, -0.08894562721252441, 0.04049629718065262, 0.13145297765731812, -0.10762684047222137, -0.00029590370832011104, 0.016895757988095284, 0.053150758147239685, -0.02384002134203911, 0.06560497730970383, 0.08113297820091248, 0.040279969573020935, -0.02908921428024769, 0.15073174238204956, 0.10429888218641281, -0.12105728685855865, 0.04198841750621796, 0.0941365659236908, -0.09897014498710632, -0.075967937707901, 0.04600611701607704, 0.06412173062562943, 0.03633765131235123, -0.025525284931063652, -0.005416040308773518, -0.06965550035238266, 0.05341847240924835, 0.06583858281373978, 0.020721372216939926, 0.017118867486715317, -0.07203074544668198, 0.005560126155614853, -0.08384118229150772, 0.10463190078735352, 0.002166942460462451, 0.03462355211377144, -0.19948604702949524, -0.03137785568833351, 0.03691473975777626, 0.025243427604436874, -0.030610833317041397, -0.03623351827263832, -0.030772026628255844, 0.011482099071145058, -0.04388830065727234, 0.0003500139282550663, -0.0037703148555010557, -0.009917370975017548, 0.011526111513376236, -0.010594231076538563, -0.03151080757379532, 0.05508873611688614, -0.03636229410767555, 0.003935730550438166, -0.09024203568696976, 0.030372576788067818, -0.0643979161977768, 0.01490132138133049, -0.02322743833065033, -0.06571394950151443, 0.054285649210214615, -0.10500659048557281, -0.021268440410494804, 0.029104875400662422, -0.10378387570381165, 0.012029694393277168, 0.051450733095407486, 0.04383710026741028, -0.019072098657488823, -0.13463912904262543, 0.061526935547590256, -0.02464192733168602, -0.0269122663885355, -0.024098476395010948, 0.11017503589391708, -0.06181040778756142, -0.0029996575322002172, -0.05195862427353859, -0.025048941373825073, -0.044009018689394, 0.10757200419902802, 0.05801590532064438, 0.07568375021219254, 0.11535059660673141, -0.02792328968644142, -0.010311901569366455, -0.1385938674211502, 0.01661893166601658, 0.005702556576579809, 0.02667849324643612, -0.09891234338283539, -0.06413057446479797, 0.027600517496466637, -0.03367125615477562, 0.04375975951552391, 0.05936301127076149, -0.038219548761844635, 0.07982878386974335, -0.052227262407541275, -0.12032372504472733, 0.06558059900999069, 0.03536567464470863, -0.018970783799886703, 0.03669200837612152, 0.014030280523002148, -0.02004558965563774, 0.0018705498659983277, 0.009102207608520985, 0.02561977505683899, 0.14565590023994446, 0.09084015339612961, 0.06628797948360443, 0.07058161497116089, -0.06455744802951813, -0.11726727336645126, 0.17459551990032196, -0.0877123475074768, 0.13039487600326538, -0.12425466626882553, 0.131888747215271, 0.14493946731090546, -0.11681957542896271, 0.03769003599882126, -0.09608054906129837, -0.04626304656267166, -0.0472734272480011, -0.31311777234077454, -0.06490244716405869, -0.09302473813295364, 0.013345025479793549, -0.030866477638483047, 0.0023239676374942064, -0.045170821249485016, 0.04409264400601387, -0.03314608335494995, 0.10701203346252441, 0.0032597596291452646, -0.011243380606174469, 0.05784890428185463, 0.016682062298059464, -0.05096999928355217, 0.010132149793207645, 0.09739057719707489, 0.08106988668441772, 0.02019771747291088, 0.03589323163032532, 0.08751186728477478, 0.03080054000020027, 0.0397915318608284, 0.03682250902056694, -0.06254797428846359, 0.001786823384463787, -0.03784019872546196, -0.0035793373826891184, 0.10746552795171738, 0.02850322425365448, -0.033371638506650925, -0.01461291965097189, 0.201374813914299, -0.02755080908536911, -0.09937946498394012, -0.13162867724895477, 0.05914503335952759, -0.022502737119793892, 0.041633643209934235, -0.02352227084338665, -0.05618168041110039, -0.02278287149965763, 0.20131875574588776, 0.13565725088119507, 0.04089624807238579, -0.05056225508451462, 0.06183210387825966, -0.018073786050081253, -0.1052674874663353, 0.12738479673862457, 0.06443776935338974, 0.13108928501605988, -0.04217437654733658, 0.09319281578063965, -0.05768614262342453, -0.01871771365404129, 0.025396062061190605, 0.02640601247549057, 0.01976810209453106, 0.029662545770406723, -0.0838618129491806, 0.08105558156967163, 0.025279689580202103, -0.2064574509859085, 0.10763151943683624, -0.034168146550655365, -0.08528993278741837, -0.028219660744071007, 0.10234179347753525, -0.07710706442594528, -0.009978045709431171, -0.03674164041876793, 0.0202531386166811, 0.1689864993095398, 0.02677268162369728, -0.011846333742141724, -0.13350839912891388, 0.053802188485860825, -0.14383724331855774, 0.22671040892601013, -0.01434856466948986, 0.020868908613920212, 0.05833740159869194, -0.044090189039707184, -0.19193236529827118, -0.007430537603795528, 0.009223710745573044, -0.04523365944623947, 0.017960770055651665, 0.16109035909175873, -0.0027846070006489754, 0.2021184265613556, 0.023093469440937042, -0.016438286751508713, -0.0039797257632017136, -0.0270231906324625, -0.006671652663499117, -0.0764879509806633, 0.12346600741147995, -0.10908254981040955, 0.12826429307460785, 0.11313086748123169, -0.02616165019571781, 0.009881656616926193, -0.1286027431488037, 0.03519096598029137, 0.011127541773021221, 0.07518888264894485, 0.06507684290409088, -0.157956063747406, 0.03230587765574455, -0.10009512305259705, 0.051758069545030594, -0.1540294885635376, -0.14580665528774261, -0.023580415174365044, 0.007769356481730938, -0.04330494627356529, 0.1074361503124237, 0.09313251823186874, 0.02047840692102909, -0.06057203188538551, -0.12690693140029907, 0.032311584800481796, 0.0798572301864624, -0.07742459326982498, -0.07558341324329376 ]
null
null
null
# **Q-Learning** Agent playing1 **Taxi-v3** This is a trained model of a **Q-Learning** agent playing **Taxi-v3** . ## Usage ```python model = load_from_hub(repo_id="basab1142/Taxi-v3-CQ", filename="q-learning.pkl") # Don't forget to check if you need to add additional attributes (is_slippery=False etc) env = gym.make(model["env_id"]) ```
{"tags": ["Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation"], "model-index": [{"name": "Taxi-v3-CQ", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "Taxi-v3", "type": "Taxi-v3"}, "metrics": [{"type": "mean_reward", "value": "7.52 +/- 2.77", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
basab1142/Taxi-v3-CQ
[ "Taxi-v3", "q-learning", "reinforcement-learning", "custom-implementation", "model-index", "region:us" ]
2024-02-10T14:37:36+00:00
[]
[]
TAGS #Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us
# Q-Learning Agent playing1 Taxi-v3 This is a trained model of a Q-Learning agent playing Taxi-v3 . ## Usage
[ "# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage" ]
[ "TAGS\n#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n", "# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage" ]
[ 32, 33 ]
[ "passage: TAGS\n#Taxi-v3 #q-learning #reinforcement-learning #custom-implementation #model-index #region-us \n# Q-Learning Agent playing1 Taxi-v3\n This is a trained model of a Q-Learning agent playing Taxi-v3 .\n\n ## Usage" ]
[ 0.048862796276807785, -0.16549694538116455, -0.005485367961227894, 0.02960980497300625, 0.1345081776380539, -0.01784728653728962, 0.11895976960659027, 0.07759871333837509, -0.07461097836494446, -0.055395450443029404, 0.1418241262435913, 0.09088201075792313, 0.055222880095243454, 0.05699880048632622, 0.09511256217956543, -0.27440664172172546, 0.048217080533504486, -0.02918700873851776, 0.05621987581253052, 0.11878681182861328, 0.0670095682144165, -0.040441032499074936, 0.061956584453582764, 0.11818158626556396, -0.1018151044845581, -0.007344264071434736, 0.035402704030275345, -0.09440053254365921, 0.17413531243801117, 0.07204403728246689, 0.12337774783372879, 0.05132639780640602, 0.179361954331398, -0.12762396037578583, 0.024310702458024025, -0.0010275895474478602, -0.10138072073459625, -0.03909514099359512, -0.012415820732712746, -0.08349097520112991, 0.03230205550789833, 0.23522862792015076, 0.07199250161647797, 0.06632792949676514, -0.17707863450050354, -0.06584878265857697, -0.04375573247671127, 0.069611094892025, 0.14951466023921967, 0.03758616745471954, -0.033800311386585236, 0.1684885323047638, -0.2564343810081482, 0.05066783353686333, 0.037275806069374084, -0.42313119769096375, 0.017119819298386574, 0.1507398933172226, 0.15090937912464142, 0.06909667700529099, -0.10573802888393402, 0.013512322679162025, 0.051325585693120956, -0.0005318621988408267, 0.024325110018253326, 0.006554204970598221, 0.15601307153701782, 0.08537693321704865, -0.1487821787595749, -0.058576688170433044, 0.17441977560520172, -0.03788546845316887, -0.02613203600049019, -0.039745692163705826, 0.0067160045728087425, -0.06427708268165588, -0.004067842848598957, -0.1777995079755783, 0.00734262028709054, 0.06666424125432968, -0.014348524622619152, 0.014901017770171165, -0.035522811114788055, -0.0966939702630043, -0.023098144680261612, -0.08592145889997482, 0.01677769608795643, -0.006319406442344189, -0.10187895596027374, 0.05002119392156601, -0.061138734221458435, 0.0014382408699020743, -0.05123179033398628, -0.15047866106033325, -0.049055423587560654, -0.03481535613536835, 0.1474713832139969, -0.0044205985032022, -0.01873963139951229, -0.03164304047822952, 0.15474793314933777, 0.049551334232091904, -0.05370146036148071, 0.05625450983643532, 0.07605006545782089, 0.23867930471897125, 0.10401605814695358, 0.10196955502033234, -0.06798075139522552, 0.10180158913135529, -0.12330973148345947, -0.08915644884109497, -0.17508824169635773, 0.11820860952138901, 0.00015364694991149008, 0.1317785084247589, -0.12023144960403442, 0.07898581773042679, -0.067511186003685, 0.013453764840960503, 0.01636839471757412, 0.0820009782910347, -0.012399360537528992, 0.10676060616970062, -0.005061192903667688, -0.06941985338926315, 0.014177112840116024, 0.05935845896601677, 0.03754841163754463, -0.038601722568273544, -0.03192409873008728, -0.05762290954589844, -0.05065649375319481, -0.10128600150346756, -0.06447898596525192, 0.018573462963104248, -0.007677143905311823, -0.1833900660276413, -0.06407523155212402, 0.00897200871258974, 0.015712225809693336, -0.03988850116729736, -0.05148044601082802, -0.15265507996082306, -0.042461175471544266, -0.015450406819581985, -0.03500641882419586, -0.06214277446269989, -0.0383245050907135, 0.046435944736003876, -0.07560601085424423, 0.013364278711378574, 0.023342855274677277, 0.05405820533633232, -0.025881100445985794, 0.06068144738674164, -0.08357544988393784, 0.09493788331747055, -0.1540430635213852, -0.03271956741809845, -0.025445878505706787, -0.041183918714523315, 0.1752462536096573, 0.06099751964211464, -0.015994304791092873, 0.15260063111782074, -0.17141541838645935, -0.058121129870414734, 0.15596486628055573, 0.008629098534584045, -0.09967197477817535, -0.003560945624485612, -0.09397093951702118, 0.1428760588169098, 0.08571921288967133, 0.2478504776954651, 0.12005335837602615, -0.22748184204101562, 0.055358242243528366, 0.12515293061733246, -0.14365963637828827, 0.10365243256092072, 0.07344598323106766, 0.005470725707709789, -0.18886831402778625, -0.06843198090791702, -0.06121627986431122, 0.1053021252155304, -0.08522345870733261, -0.0776243582367897, 0.09323626756668091, -0.05086790770292282, 0.24641476571559906, -0.028281206265091896, 0.06174173951148987, -0.026681531220674515, -0.1389324963092804, -0.01723906397819519, 0.060955192893743515, 0.05258452147245407, -0.024835573509335518, -0.25895482301712036, 0.13646544516086578, 0.048650871962308884, 0.025074828416109085, 0.004106190986931324, -0.05691491439938545, 0.016934165731072426, 0.1511998474597931, 0.020012924447655678, 0.13717477023601532, 0.027723990380764008, 0.0706823319196701, -0.006239562761038542, -0.10560829937458038, -0.04169593006372452, 0.061916545033454895, -0.08518962562084198, -0.06641357392072678, 0.011197872459888458, -0.06935211271047592, -0.11783787608146667, -0.12166737765073776, -0.026334572583436966, -0.02980303019285202, -0.07444227486848831, 0.02368103712797165, 0.06536602973937988, -0.06702698022127151, -0.0023908785078674555, 0.007125476840883493, -0.011537045240402222, 0.16434046626091003, 0.011393417604267597, -0.007796820718795061, 0.1328643560409546, -0.11533161997795105, 0.12461213022470474, 0.049438029527664185, -0.024806302040815353, -0.04662557691335678, 0.0014137453399598598, -0.057529181241989136, 0.029044216498732567, -0.04390640929341316, 0.02774495631456375, 0.20111067593097687, 0.02772962674498558, 0.11389166116714478, -0.0656520202755928, 0.04385066404938698, -0.007961965166032314, -0.009693224914371967, 0.018563594669103622, 0.07608018070459366, 0.07813210040330887, -0.1324140727519989, 0.02262016013264656, 0.22455167770385742, 0.1385764330625534, 0.18313980102539062, -0.010877152904868126, 0.06325667351484299, -0.04875868931412697, 0.027505528181791306, 0.024100203067064285, 0.10314226150512695, -0.10732068121433258, -0.0322517491877079, -0.025407759472727776, 0.023599207401275635, -0.08197105675935745, -0.1055799350142479, -0.090115025639534, 0.01222382951527834, -0.03125503659248352, -0.15570329129695892, 0.13300658762454987, -0.10451057553291321, 0.01802753657102585, 0.04692702740430832, -0.22163605690002441, 0.11530312895774841, 0.014291439205408096, -0.10303618758916855, 0.11281087249517441, -0.12051989883184433, -0.08699832111597061, -0.05777236074209213, -0.18658851087093353, 0.05280197039246559, 0.04673841595649719, 0.05166793242096901, -0.18521739542484283, 0.024835903197526932, 0.05545609071850777, 0.13426995277404785, -0.09743253141641617, -0.07142634689807892, -0.15038461983203888, 0.016068490222096443, -0.033661190420389175, -0.16029728949069977, -0.005609163548797369, -0.032781440764665604, -0.18849676847457886, -0.04539939761161804, -0.15086813271045685, -0.034627582877874374, 0.20464378595352173, 0.026907702907919884, 0.09480511397123337, -0.07926445454359055, 0.3802889585494995, -0.042039383202791214, -0.06146497279405594, -0.01321389526128769, -0.07072482258081436, 0.02512686513364315, 0.13271741569042206, 0.0036099457647651434, -0.017886579036712646, -0.0037857077550143003, 0.0024592927657067776, -0.06234965845942497, -0.13400450348854065, 0.0028710351325571537, 0.03905198723077774, 0.1874423623085022, 0.004639793653041124, 0.06659388542175293, 0.03133883699774742, 0.057546284049749374, 0.07748064398765564, 0.030926106497645378, 0.0011591583024710417, -0.01591806672513485, 0.06604493409395218, -0.11684755235910416, 0.042466625571250916, -0.030429253354668617, -0.10143838077783585, -0.013183288276195526, 0.07950251549482346, 0.12755028903484344, 0.17849206924438477, -0.04790908098220825, 0.17489230632781982, 0.13580141961574554, 0.16576050221920013, 0.049315933138132095, -0.020801831036806107, -0.08773037046194077, -0.06118565797805786, 0.004774159751832485, -0.031952597200870514, 0.04869702458381653, 0.3231290578842163, 0.037619613111019135, -0.09036035090684891, 0.11149907857179642, 0.009480619803071022, 0.05359881371259689, 0.022797370329499245, -0.11162138730287552, 0.11170321702957153, 0.07968773692846298, -0.06341761350631714, -0.07602835446596146, 0.16758501529693604, -0.1109386757016182, -0.26646625995635986, -0.11410990357398987, -0.012305386364459991, 0.07903840392827988, 0.005651174578815699, 0.05498376116156578, -0.11829282343387604, -0.16034497320652008, -0.034191906452178955, 0.1335442066192627, -0.3077351450920105, 0.2065143585205078, -0.0198091771453619, 0.06707923114299774, -0.039657969027757645, -0.07026876509189606, 0.09694647043943405, 0.13174086809158325, 0.29124146699905396, 0.01396956667304039, 0.04841272905468941, -0.15176129341125488, -0.0976925864815712, 0.0018439020495861769, 0.015482662245631218, -0.02563396655023098, 0.028520405292510986, -0.0540912002325058, 0.008404579944908619, -0.018086453899741173, 0.2102297693490982, -0.11316607892513275, 0.004344627261161804, -0.06968966871500015, -0.11707738786935806, 0.19409789144992828, -0.07178345322608948, -0.04543264955282211, -0.14959357678890228, -0.15512511134147644, -0.004174166824668646, -0.02413962036371231, -0.019664527848362923, -0.17603960633277893, -0.18804074823856354, -0.05204557999968529, -0.005645004566758871, -0.003464865731075406, 0.05867868289351463, -0.07517234236001968, -0.04805335775017738, 0.1009904220700264, -0.07743175327777863, -0.056063808500766754, -0.1103200614452362, 0.1391381323337555, 0.06248528137803078, 0.16743235290050507, 0.05907081440091133, 0.0006117874872870743, 0.11471151560544968, -0.02913086675107479, 0.11103474348783493, -0.11291708797216415, -0.17145049571990967, -0.08334989100694656, -0.018775060772895813, 0.09519003331661224, -0.04789286106824875, 0.0028788831550627947, 0.2550160884857178, 0.14880181849002838, -0.0897710770368576, 0.27680760622024536, 0.04414956644177437, -0.09375058114528656, -0.18432219326496124, -0.15961645543575287, 0.03759992495179176, 0.060025621205568314, 0.13095876574516296, -0.057205069810152054, -0.08483537286520004, -0.08492398262023926, -0.07478608191013336, -0.13140805065631866, -0.24232175946235657, -0.030598774552345276, 0.22874866425991058, 0.08656918257474899, 0.08219650387763977, -0.012482990510761738, -0.01186054851859808, 0.00526038184762001, 0.02680150233209133, 0.12018456310033798, -0.13341329991817474, 0.11107480525970459, 0.022198403254151344, 0.044267985969781876, 0.009712530300021172, 0.07929777354001999, 0.03375575691461563, -0.003218587953597307, -0.0006439819699153304, -0.0988350659608841, -0.2596651017665863, 0.0816885456442833, -0.01623627357184887, -0.09960969537496567, 0.014988959766924381, 0.02061903104186058, -0.2089255303144455, 0.011128270998597145, -0.019883770495653152, -0.03150356933474541, -0.06483490765094757, -0.10664787143468857, -0.056551624089479446, 0.04928823933005333, 0.10853826254606247, 0.011660109274089336, 0.05354316532611847, -0.0404130220413208, 0.07917837053537369, 0.0826287642121315, 0.15132710337638855, 0.06795957684516907, -0.190711110830307, -0.10953907668590546, -0.0414445661008358, 0.12121522426605225, -0.12505418062210083, 0.036917757242918015, 0.053161121904850006, -0.016534561291337013, 0.14621229469776154, 0.1070784479379654, -0.07452095299959183, 0.11915595084428787, 0.08904775977134705, -0.04094788804650307, -0.23367151618003845, -0.07120766490697861, 0.11133213341236115, 0.07195597887039185, -0.03961895406246185, 0.018120890483260155, -0.04960581287741661, -0.013980977237224579, 0.048759616911411285, -0.0538676381111145, -0.07230538129806519, 0.004421027842909098, 0.1247575581073761, 0.1029362753033638, -0.04655474051833153, 0.01296416949480772, 0.037371400743722916, 0.003788623260334134, 0.04730486497282982, 0.0407949760556221, -0.08269952982664108, -0.04124005511403084, 0.02782733179628849, 0.37552911043167114, -0.010165480896830559, -0.020456433296203613, 0.018555615097284317, -0.19949445128440857, 0.09135842323303223, 0.13205479085445404, 0.04697350412607193, 0.004247748292982578, -0.08139242231845856, 0.026877427473664284, -0.010625290684401989, 0.09936143457889557, -0.07806670665740967, -0.05493134260177612, -0.21631066501140594, -0.025010565295815468, 0.017490221187472343, 0.24077683687210083, -0.08458559215068817, -0.12801732122898102, -0.20628872513771057, 0.13128381967544556, -0.11333390325307846, -0.03695881739258766, -0.024473199620842934, 0.03926658630371094, -0.01989821158349514, 0.06291737407445908, -0.0710630789399147, 0.006373001262545586, -0.11024709790945053, 0.055267609655857086, 0.04204455390572548, 0.1229788213968277, 0.014207782223820686, 0.02016810141503811, 0.05822525918483734, -0.01837925612926483, 0.07173580676317215, -0.06203491613268852, -0.04550490900874138, 0.14224006235599518, -0.020255116745829582, -0.04152837023139, -0.0483345128595829, -0.036874305456876755, 0.11981741338968277, -0.05059147998690605, -0.007141099311411381, -0.054929375648498535, -0.06906463205814362, 0.03462086617946625, -0.009175732731819153, -0.008798843249678612, 0.06801853328943253, 0.04024988040328026, -0.026994358748197556, 0.005263668950647116, 0.03447828069329262, -0.10330043733119965, -0.04955084249377251, 0.16955432295799255, -0.0749620869755745, 0.10274054110050201, -0.031069839373230934, 0.018015999346971512, 0.005847334861755371, -0.022399673238396645, -0.015360680408775806, -0.1457086056470871, -0.06137600541114807, -0.09489979594945908, 0.11565322428941727, 0.08146517723798752, 0.03358805552124977, 0.04274565726518631, 0.019532648846507072, -0.04414922371506691, -0.038583990186452866, 0.12961317598819733, 0.08133101463317871, 0.012996876612305641, 0.01137041300535202, 0.01941833831369877, -0.020302120596170425, 0.0028480992186814547, -0.01250747125595808, -0.07239153981208801, -0.05874783173203468, 0.09400010108947754, 0.1600283533334732, -0.06127211079001427, -0.13325586915016174, -0.020593497902154922, 0.04988488554954529, 0.0014717020094394684, -0.08777432143688202, 0.04833676666021347, 0.15805292129516602, -0.05623878911137581, 0.03216489031910896, -0.09984751045703888, -0.07263360917568207, -0.16060975193977356, -0.10029061883687973, -0.06092562898993492, -0.28350353240966797, 0.09752398729324341, 0.006392303854227066, -0.014731393195688725, 0.059529416263103485, 0.051305368542671204, -0.052508849650621414, 0.07068239152431488, -0.18146829307079315, -0.007054794579744339, 0.03497592359781265, -0.13212306797504425, 0.02475893869996071, -0.2378365397453308, 0.10198072344064713, -0.04623803123831749, -0.1519704908132553, -0.04004510119557381, 0.0641569048166275, -0.09540136158466339, -0.01822364516556263, -0.0475153923034668, -0.01922670193016529, 0.01624443754553795, -0.009348669089376926, -0.031147832050919533, 0.13716529309749603, 0.02827494591474533, -0.03268734738230705, 0.005254602525383234, 0.0223685409873724, 0.03955082967877388, -0.0969657450914383, -0.05986930429935455, 0.08311155438423157, -0.031056145206093788, 0.14728976786136627, 0.000341245875461027, 0.04181376099586487, -0.06758682429790497, 0.2593761384487152, 0.2023983597755432, -0.12479214370250702, 0.008118697442114353, -0.021801479160785675, 0.012670028023421764, -0.041751839220523834, 0.13110700249671936, 0.013386172242462635, 0.12186761200428009, -0.17513342201709747, -0.01036517322063446, -0.0818324014544487, -0.04501292482018471, 0.06702108681201935, 0.14714950323104858, 0.15742522478103638, 0.03436789661645889, -0.07328428328037262, 0.06722653657197952, -0.30119743943214417, 0.20540550351142883, -0.1346001923084259, -0.01498429011553526, -0.040251150727272034, -0.058389630168676376, 0.061147745698690414, 0.11309876292943954, 0.10832664370536804, -0.021150551736354828, -0.0905047357082367, -0.04486766457557678, -0.039378076791763306, -0.13019338250160217, -0.02718670479953289, 0.1654091775417328, 0.06799814850091934, 0.31520840525627136, -0.017577875405550003, 0.07702425122261047, 0.034410297870635986, 0.06451138854026794, 0.004519328009337187, 0.09537279605865479, 0.07960964739322662, -0.06345855444669724, -0.07373003661632538, -0.001637450186535716, 0.05033271387219429, 0.14567798376083374, -0.03826142102479935, -0.18691548705101013, 0.15858715772628784, 0.07192251086235046, -0.13762691617012024, -0.05777517706155777, 0.08409425616264343, -0.0739973932504654, 0.0550808347761631, 0.08115427941083908, 0.015876613557338715, -0.017793258652091026, -0.004664506763219833, 0.06074233725667, 0.024694660678505898, -0.02343848906457424, 0.003570882137864828, -0.08337053656578064, -0.04151543974876404, 0.07267895340919495, -0.0844460055232048, -0.20546193420886993, -0.0957019031047821, -0.07551700621843338, 0.030557552352547646, -0.0649830624461174, 0.12575586140155792, 0.1717868149280548, 0.0593598335981369, -0.03307248651981354, -0.10721943527460098, -0.035562749952077866, 0.07602505385875702, -0.044773899018764496, -0.09409699589014053 ]
null
null
null
<br> # Vim Model Card ## Model Details Vision Mamba (Vim) is a generic backbone trained on the ImageNet-1K dataset for vision tasks. - **Developed by:** [HUST](https://english.hust.edu.cn/), [Horizon Robotics](https://en.horizon.cc/), [BAAI](https://www.baai.ac.cn/english.html) - **Model type:** A generic vision backbone based on the bidirectional state space model (SSM) architecture. - **License:** Non-commercial license ### Model Sources - **Repository:** https://github.com/hustvl/Vim - **Paper:** https://arxiv.org/abs/2401.09417 ## Uses The primary use of Vim is research on vision tasks, e.g., classification, segmentation, detection, and instance segmentation, with an SSM-based backbone. The primary intended users of the model are researchers and hobbyists in computer vision, machine learning, and artificial intelligence. ## How to Get Started with the Model - You can replace the backbone for vision tasks with the proposed Vim: https://github.com/hustvl/Vim/blob/main/vim/models_mamba.py - Then you can load this checkpoint and start training. ## Training Details Vim is pretrained on ImageNet-1K with classification supervision. The training data is around 1.3M images from [ImageNet-1K dataset](https://www.image-net.org/challenges/LSVRC/2012/). See more details in this [paper](https://arxiv.org/abs/2401.09417). ## Evaluation Vim-tiny is evaluated on ImageNet-1K val set, and achieves 76.1% Top-1 Acc. By further finetuning at finer granularity, Vim-tiny achieves 78.3% Top-1 Acc. See more details in this [paper](https://arxiv.org/abs/2401.09417). ## Additional Information ### Citation Information ``` @article{vim, title={Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model}, author={Lianghui Zhu and Bencheng Liao and Qian Zhang and Xinlong Wang and Wenyu Liu and Xinggang Wang}, journal={arXiv preprint arXiv:2401.09417}, year={2024} } ```
{"license": "apache-2.0"}
null
hustvl/Vim-tiny-midclstok
[ "arxiv:2401.09417", "license:apache-2.0", "region:us" ]
2024-02-10T14:40:18+00:00
[ "2401.09417" ]
[]
TAGS #arxiv-2401.09417 #license-apache-2.0 #region-us
<br> # Vim Model Card ## Model Details Vision Mamba (Vim) is a generic backbone trained on the ImageNet-1K dataset for vision tasks. - Developed by: HUST, Horizon Robotics, BAAI - Model type: A generic vision backbone based on the bidirectional state space model (SSM) architecture. - License: Non-commercial license ### Model Sources - Repository: URL - Paper: URL ## Uses The primary use of Vim is research on vision tasks, e.g., classification, segmentation, detection, and instance segmentation, with an SSM-based backbone. The primary intended users of the model are researchers and hobbyists in computer vision, machine learning, and artificial intelligence. ## How to Get Started with the Model - You can replace the backbone for vision tasks with the proposed Vim: URL - Then you can load this checkpoint and start training. ## Training Details Vim is pretrained on ImageNet-1K with classification supervision. The training data is around 1.3M images from ImageNet-1K dataset. See more details in this paper. ## Evaluation Vim-tiny is evaluated on ImageNet-1K val set, and achieves 76.1% Top-1 Acc. By further finetuning at finer granularity, Vim-tiny achieves 78.3% Top-1 Acc. See more details in this paper. ## Additional Information
[ "# Vim Model Card", "## Model Details\n\nVision Mamba (Vim) is a generic backbone trained on the ImageNet-1K dataset for vision tasks.\n\n- Developed by: HUST, Horizon Robotics, BAAI\n- Model type: A generic vision backbone based on the bidirectional state space model (SSM) architecture.\n- License: Non-commercial license", "### Model Sources\n\n- Repository: URL\n- Paper: URL", "## Uses\n\nThe primary use of Vim is research on vision tasks, e.g., classification, segmentation, detection, and instance segmentation, with an SSM-based backbone.\nThe primary intended users of the model are researchers and hobbyists in computer vision, machine learning, and artificial intelligence.", "## How to Get Started with the Model\n\n- You can replace the backbone for vision tasks with the proposed Vim: URL\n- Then you can load this checkpoint and start training.", "## Training Details\n\nVim is pretrained on ImageNet-1K with classification supervision.\nThe training data is around 1.3M images from ImageNet-1K dataset.\nSee more details in this paper.", "## Evaluation\n\nVim-tiny is evaluated on ImageNet-1K val set, and achieves 76.1% Top-1 Acc. By further finetuning at finer granularity, Vim-tiny achieves 78.3% Top-1 Acc. See more details in this paper.", "## Additional Information" ]
[ "TAGS\n#arxiv-2401.09417 #license-apache-2.0 #region-us \n", "# Vim Model Card", "## Model Details\n\nVision Mamba (Vim) is a generic backbone trained on the ImageNet-1K dataset for vision tasks.\n\n- Developed by: HUST, Horizon Robotics, BAAI\n- Model type: A generic vision backbone based on the bidirectional state space model (SSM) architecture.\n- License: Non-commercial license", "### Model Sources\n\n- Repository: URL\n- Paper: URL", "## Uses\n\nThe primary use of Vim is research on vision tasks, e.g., classification, segmentation, detection, and instance segmentation, with an SSM-based backbone.\nThe primary intended users of the model are researchers and hobbyists in computer vision, machine learning, and artificial intelligence.", "## How to Get Started with the Model\n\n- You can replace the backbone for vision tasks with the proposed Vim: URL\n- Then you can load this checkpoint and start training.", "## Training Details\n\nVim is pretrained on ImageNet-1K with classification supervision.\nThe training data is around 1.3M images from ImageNet-1K dataset.\nSee more details in this paper.", "## Evaluation\n\nVim-tiny is evaluated on ImageNet-1K val set, and achieves 76.1% Top-1 Acc. By further finetuning at finer granularity, Vim-tiny achieves 78.3% Top-1 Acc. See more details in this paper.", "## Additional Information" ]
[ 23, 5, 78, 15, 69, 40, 43, 60, 5 ]
[ "passage: TAGS\n#arxiv-2401.09417 #license-apache-2.0 #region-us \n# Vim Model Card## Model Details\n\nVision Mamba (Vim) is a generic backbone trained on the ImageNet-1K dataset for vision tasks.\n\n- Developed by: HUST, Horizon Robotics, BAAI\n- Model type: A generic vision backbone based on the bidirectional state space model (SSM) architecture.\n- License: Non-commercial license### Model Sources\n\n- Repository: URL\n- Paper: URL## Uses\n\nThe primary use of Vim is research on vision tasks, e.g., classification, segmentation, detection, and instance segmentation, with an SSM-based backbone.\nThe primary intended users of the model are researchers and hobbyists in computer vision, machine learning, and artificial intelligence.## How to Get Started with the Model\n\n- You can replace the backbone for vision tasks with the proposed Vim: URL\n- Then you can load this checkpoint and start training.## Training Details\n\nVim is pretrained on ImageNet-1K with classification supervision.\nThe training data is around 1.3M images from ImageNet-1K dataset.\nSee more details in this paper.## Evaluation\n\nVim-tiny is evaluated on ImageNet-1K val set, and achieves 76.1% Top-1 Acc. By further finetuning at finer granularity, Vim-tiny achieves 78.3% Top-1 Acc. See more details in this paper.## Additional Information" ]
[ -0.06916045397520065, 0.07929331064224243, -0.004483504686504602, 0.006933827418833971, 0.10427270829677582, 0.02840515598654747, 0.08510330319404602, 0.0942901223897934, -0.02999553270637989, 0.026321377605199814, 0.05504373833537102, 0.026886777952313423, 0.11002620309591293, 0.17623749375343323, 0.09143801033496857, -0.18970540165901184, 0.013574239797890186, -0.03552014008164406, 0.03781726956367493, 0.04518246650695801, 0.12129514664411545, -0.08707787096500397, 0.0475257970392704, -0.010393248870968819, -0.12182798236608505, 0.02520960569381714, 0.003757215104997158, 0.01696396805346012, 0.13680942356586456, -0.006761287804692984, 0.07933206111192703, -0.02776586264371872, 0.10129541903734207, -0.13574907183647156, 0.026744043454527855, 0.09777875989675522, 0.04597890377044678, 0.06037607043981552, 0.05494625121355057, 0.13592788577079773, 0.1434420943260193, -0.05881083756685257, 0.0313713438808918, 0.03867887333035469, -0.022097768262028694, -0.10790886729955673, -0.11517560482025146, 0.07486101984977722, 0.10595742613077164, 0.05711532384157181, -0.006067706272006035, 0.11858930438756943, 0.050723519176244736, 0.06021462380886078, 0.0010873852297663689, -0.258524090051651, -0.061712104827165604, 0.18962015211582184, -0.004436175338923931, 0.053296707570552826, -0.05090909078717232, 0.04391380026936531, -0.009644020348787308, 0.025029223412275314, 0.05560844764113426, 0.04045868292450905, -0.04314357787370682, -0.0514335073530674, -0.13920584321022034, -0.04930907487869263, 0.1325898915529251, 0.05651530995965004, -0.050628628581762314, -0.09743232280015945, -0.07365100085735321, 0.02765483222901821, 0.019454320892691612, -0.03582943230867386, 0.021824240684509277, 0.028774544596672058, 0.1280982792377472, -0.08087906986474991, -0.12235275655984879, -0.02511211857199669, 0.03315041586756706, -0.07077526301145554, 0.04440613090991974, 0.06610046327114105, 0.04015855863690376, 0.09084262698888779, -0.10266314446926117, -0.07240545749664307, -0.024843694642186165, -0.05719388648867607, -0.0800919234752655, -0.05759824067354202, 0.03476304933428764, -0.22716262936592102, -0.03754924237728119, 0.04282180592417717, 0.0691986009478569, 0.07348783314228058, -0.038193732500076294, 0.04501540958881378, 0.02570069581270218, 0.10178631544113159, -0.09381885081529617, 0.026167122647166252, 0.06536070257425308, 0.02621140331029892, 0.0340069942176342, -0.003545308019965887, -0.022761112079024315, 0.005064570810645819, 0.02348867617547512, 0.024495430290699005, -0.014853378757834435, -0.0009883464081212878, 0.035308998078107834, -0.04178423807024956, 0.11989831179380417, -0.10982534289360046, 0.015194306150078773, 0.00400315597653389, -0.022269733250141144, 0.015260383486747742, 0.05078543722629547, -0.02322303131222725, -0.10549471527338028, -0.013128574937582016, -0.009505373425781727, -0.04071176424622536, -0.09980783611536026, -0.08699009567499161, 0.059242166578769684, -0.15353454649448395, -0.06353821605443954, -0.11232036352157593, -0.06724347174167633, -0.058450933545827866, 0.1482870727777481, -0.0759664922952652, -0.024909142404794693, -0.02473381534218788, -0.05318816006183624, 0.01317769754678011, 0.0579933300614357, 0.12626713514328003, 0.011998958885669708, 0.0037273571360856295, -0.047519028186798096, 0.08949407935142517, 0.05296152085065842, -0.00041690253419801593, -0.005623442120850086, 0.005619698204100132, -0.11204876005649567, 0.015348203480243683, -0.05177529901266098, 0.03846367448568344, -0.17207498848438263, 0.02334754168987274, -0.04200345650315285, -0.017110129818320274, 0.08975256234407425, 0.0874730721116066, -0.2044384926557541, 0.0415654331445694, 0.04919141158461571, -0.1319488137960434, -0.07612933963537216, 0.03368163853883743, 0.006186456419527531, 0.12436460703611374, 0.02015192061662674, 0.1032896637916565, 0.07053577899932861, -0.1650247871875763, -0.023597080260515213, 0.009219173341989517, -0.05502409487962723, -0.08203478902578354, 0.04132748022675514, 0.10157762467861176, -0.03292585164308548, 0.005237852223217487, -0.13009946048259735, 0.01807902194559574, -0.024668075144290924, -0.0818357914686203, -0.04282130300998688, -0.03642003983259201, -0.09980083256959915, 0.022494381293654442, 0.03404448553919792, 0.03794998675584793, -0.06106669455766678, -0.05901937931776047, 0.08687271922826767, -0.0167111586779356, 0.015202326700091362, -0.1001279428601265, 0.014296779409050941, -0.14585401117801666, 0.05817278102040291, -0.127765491604805, 0.024454817175865173, 0.06403586268424988, -0.0305617768317461, 0.08213968575000763, 0.05004265531897545, 0.03607850894331932, 0.006335393991321325, -0.038699254393577576, -0.0469917431473732, -0.008087480440735817, -0.027255862951278687, -0.030828891322016716, -0.19859910011291504, -0.06512940675020218, -0.05306384712457657, -0.052667852491140366, -0.16227935254573822, 0.00009381925337947905, 0.012114287354052067, 0.11960920691490173, 0.03774426504969597, -0.061912652105093, 0.016807761043310165, -0.03177223354578018, -0.004713296424597502, -0.055619340389966965, 0.05213990435004234, 0.03494662791490555, 0.05124911293387413, 0.01169724389910698, 0.028077583760023117, -0.11846299469470978, 0.056831635534763336, -0.0339161641895771, -0.0666639655828476, 0.08682820200920105, -0.016918979585170746, -0.01788620837032795, -0.06482874602079391, 0.036649636924266815, 0.01333752553910017, 0.05388130620121956, 0.07327370345592499, -0.049434781074523926, -0.019264021888375282, 0.05241924896836281, 0.011658410541713238, -0.07621386647224426, 0.027443891391158104, 0.22811251878738403, -0.1954983025789261, -0.030146658420562744, 0.009183586575090885, 0.0001091738959075883, 0.06153147295117378, 0.006222348194569349, -0.05325198546051979, -0.04083342105150223, 0.035366762429475784, 0.020643578842282295, 0.15914179384708405, -0.1315244734287262, -0.015086974017322063, 0.01888064108788967, 0.016117405146360397, 0.008197015151381493, -0.13184253871440887, 0.027400804683566093, 0.006862776353955269, -0.0033327059354633093, 0.028423836454749107, 0.00956409890204668, -0.07686310261487961, 0.05836877226829529, 0.09036262333393097, 0.04184629023075104, -0.050696369260549545, -0.0653306245803833, -0.12733183801174164, 0.12030946463346481, -0.07942219078540802, -0.23166389763355255, -0.15903730690479279, 0.021038437262177467, -0.026756390929222107, -0.01055446919053793, -0.028986504301428795, -0.11318889260292053, -0.09386028349399567, -0.11081454902887344, 0.006882872432470322, 0.0021654879674315453, 0.007227790541946888, 0.04181267321109772, 0.05698259547352791, 0.05920395255088806, -0.10999477654695511, 0.040450770407915115, -0.0018626609817147255, -0.017914049327373505, 0.049456242471933365, 0.054464180022478104, 0.022459236904978752, 0.14738550782203674, -0.040115054696798325, 0.06262431293725967, 0.005416065454483032, 0.24799010157585144, -0.06220084801316261, 0.14153043925762177, 0.18109017610549927, 0.02304563671350479, 0.010713365860283375, 0.20106294751167297, 0.028315046802163124, -0.06510361284017563, 0.025953233242034912, 0.017116475850343704, -0.03834326192736626, -0.21583552658557892, -0.0006604355876334012, -0.044554803520441055, -0.09000680595636368, 0.08949319273233414, 0.09780468046665192, -0.04102515056729317, 0.037758879363536835, 0.00091398018412292, -0.013270490802824497, 0.028940143063664436, 0.09010349214076996, 0.12411793321371078, -0.04041116684675217, 0.02588173747062683, -0.11094769835472107, 0.07911751419305801, 0.0356026291847229, 0.11214354634284973, 0.23407231271266937, 0.016294417902827263, 0.016076138243079185, 0.10014866292476654, 0.06607790291309357, -0.017757423222064972, 0.05387445166707039, -0.1031542494893074, -0.0016952354926615953, -0.00858671497553587, -0.07634834200143814, 0.020346693694591522, 0.06390677392482758, -0.13104519248008728, -0.044797349721193314, 0.00346628250554204, 0.057682376354932785, 0.004062701482325792, 0.1954682320356369, 0.0896979346871376, -0.14754655957221985, -0.03668694943189621, 0.0036499006673693657, -0.07533840835094452, -0.11526761949062347, -0.05045032128691673, 0.10146239399909973, -0.02637260966002941, 0.039078522473573685, -0.04130154475569725, 0.07825689762830734, -0.053795650601387024, -0.04299705848097801, 0.07281436026096344, 0.034782905131578445, 0.04458504915237427, 0.06387539952993393, -0.039090704172849655, 0.1431727260351181, 0.005266401916742325, 0.1569528877735138, -0.03701150044798851, 0.02261270582675934, -0.04377808794379234, 0.1263131946325302, 0.11972852051258087, 0.029736217111349106, -0.02895020693540573, -0.01143020298331976, 0.005033252295106649, 0.017056401818990707, 0.04660353064537048, 0.013000181876122952, 0.07588838785886765, -0.03376312181353569, -0.03232339769601822, 0.006210102699697018, -0.035984497517347336, -0.17114804685115814, -0.10321155190467834, 0.020162254571914673, -0.005245010368525982, -0.024218177422881126, -0.08881140500307083, -0.053818970918655396, 0.01653881184756756, 0.19930078089237213, -0.12903518974781036, -0.00724844029173255, -0.09260497242212296, 0.047223061323165894, 0.04393373802304268, -0.027631735429167747, 0.07604671269655228, -0.05843541771173477, 0.18044549226760864, -0.02371990866959095, -0.09528779238462448, 0.0067267753183841705, -0.05901416763663292, -0.06986793875694275, -0.04297908395528793, 0.08085035532712936, -0.00815401878207922, 0.004816103260964155, 0.01966044120490551, 0.03327743709087372, 0.028637921437621117, -0.02883259207010269, 0.05677156150341034, 0.1596723198890686, -0.07967590540647507, -0.09019174426794052, -0.054343875497579575, -0.06349363923072815, -0.03382383659482002, -0.10373940318822861, 0.06232545152306557, 0.1919141411781311, -0.06282206624746323, 0.1220935508608818, 0.10461446642875671, -0.1273232251405716, -0.26248031854629517, 0.05750667303800583, 0.05816129222512245, 0.05959125608205795, -0.058747805655002594, -0.22358667850494385, 0.07957220822572708, -0.09896083176136017, -0.03835676237940788, 0.0119758490473032, -0.23622368276119232, -0.09813948720693588, -0.042678091675043106, 0.14292798936367035, 0.05494768172502518, -0.00043313493370078504, -0.0160809438675642, 0.0056297313421964645, -0.009899682365357876, 0.06449632346630096, 0.09001680463552475, 0.09874504059553146, -0.0018609982216730714, -0.03316265717148781, 0.006462156772613525, -0.05266069248318672, 0.057304441928863525, -0.07976473122835159, 0.0315251387655735, -0.02386963739991188, 0.024260398000478745, 0.08345769345760345, -0.07188825309276581, 0.14849330484867096, 0.10081874579191208, 0.026181047782301903, -0.11014998704195023, -0.04990153759717941, -0.055528342723846436, 0.006239261478185654, -0.03686182200908661, -0.028569471091032028, -0.04469536617398262, 0.09106115251779556, 0.048059508204460144, -0.032548677176237106, 0.06066769361495972, 0.0034391721710562706, -0.0695628747344017, 0.14979827404022217, 0.0630924329161644, -0.10348392277956009, 0.025116441771388054, 0.06202254816889763, -0.016727019101381302, 0.08914326876401901, -0.22471578419208527, 0.004800311289727688, 0.08348243683576584, 0.003243420971557498, 0.03756871446967125, -0.0002798958448693156, -0.13442957401275635, -0.06979599595069885, 0.0963997095823288, -0.03818843141198158, -0.18517662584781647, -0.0015509102959185839, 0.1524362862110138, -0.08590200543403625, 0.03946267068386078, 0.13432860374450684, -0.11252912878990173, 0.0002253663697047159, 0.017279163002967834, 0.05037525296211243, -0.02420595847070217, 0.06153909116983414, 0.08407635241746902, 0.0422540009021759, -0.028486331924796104, 0.14638225734233856, 0.10908754914999008, -0.1221175268292427, 0.04318380355834961, 0.1043182834982872, -0.09875751286745071, -0.07735051959753036, 0.054611966013908386, 0.05812748894095421, 0.03380636125802994, -0.020695216953754425, -0.001297511043958366, -0.0692763403058052, 0.06043794006109238, 0.07347992807626724, 0.02069389447569847, 0.015410657040774822, -0.07255727797746658, 0.005317542236298323, -0.08847152441740036, 0.09876161813735962, -0.005605987273156643, 0.03672294318675995, -0.20008331537246704, -0.03286941349506378, 0.0352647639811039, 0.028665605932474136, -0.03085371106863022, -0.03971313685178757, -0.03194034844636917, 0.012952529825270176, -0.029750417917966843, 0.007611417677253485, -0.002070612506940961, -0.011508112773299217, 0.014118104241788387, -0.011249038390815258, -0.03101459890604019, 0.05800807476043701, -0.03589310497045517, 0.0011717402376234531, -0.09725215286016464, 0.026973124593496323, -0.06606727838516235, 0.01412163581699133, -0.026022259145975113, -0.06907548010349274, 0.051730379462242126, -0.10983260720968246, -0.026307296007871628, 0.023979006335139275, -0.10016471892595291, 0.013169034384191036, 0.050865188241004944, 0.041400033980607986, -0.022836241871118546, -0.13263830542564392, 0.05829960107803345, -0.025303101167082787, -0.02805565856397152, -0.020856205374002457, 0.11770911514759064, -0.06296113133430481, -0.0029806317761540413, -0.05683073401451111, -0.030463075265288353, -0.038808539509773254, 0.11035477370023727, 0.0608409158885479, 0.07415973395109177, 0.11640825867652893, -0.03232822194695473, -0.005546851083636284, -0.13783982396125793, 0.016258100047707558, 0.010609517805278301, 0.031500622630119324, -0.0919376090168953, -0.06994550675153732, 0.027075225487351418, -0.02851938270032406, 0.030786653980612755, 0.05955471470952034, -0.04176202043890953, 0.08128218352794647, -0.05627487599849701, -0.13093331456184387, 0.06311988085508347, 0.03607342764735222, -0.01570625975728035, 0.030311694368720055, 0.016016779467463493, -0.027245378121733665, 0.0029121190309524536, 0.013977193273603916, 0.028053712099790573, 0.14415210485458374, 0.0907992571592331, 0.06499844789505005, 0.0677289143204689, -0.06575886160135269, -0.11759163439273834, 0.16862556338310242, -0.08714838325977325, 0.12706875801086426, -0.12087594717741013, 0.13354501128196716, 0.1422709822654724, -0.10884124040603638, 0.037899117916822433, -0.09513013064861298, -0.04607725888490677, -0.04790712893009186, -0.3084564507007599, -0.06774155050516129, -0.0913812592625618, 0.010927368886768818, -0.02710932493209839, 0.004081621300429106, -0.04313855990767479, 0.04843761771917343, -0.02958184853196144, 0.10958312451839447, 0.01558868307620287, -0.010856765322387218, 0.056579068303108215, 0.012079409323632717, -0.05207757651805878, 0.0136796860024333, 0.1015552207827568, 0.08077817410230637, 0.018186435103416443, 0.037185244262218475, 0.09543709456920624, 0.03154018521308899, 0.040231551975011826, 0.03522360324859619, -0.06587717682123184, 0.004607714246958494, -0.038566671311855316, -0.0005406649434007704, 0.11751437932252884, 0.02808183804154396, -0.031283311545848846, -0.011132978834211826, 0.19929473102092743, -0.023619184270501137, -0.10718797892332077, -0.1303643435239792, 0.06368143111467361, -0.024378081783652306, 0.03723849728703499, -0.02631954848766327, -0.055556684732437134, -0.03145613148808479, 0.2074611932039261, 0.1369287520647049, 0.05084270238876343, -0.04924369603395462, 0.06792665272951126, -0.017610754817724228, -0.10042189806699753, 0.1328490972518921, 0.05994858220219612, 0.13258764147758484, -0.04888513684272766, 0.0907634049654007, -0.06271868199110031, -0.017073102295398712, 0.018423806875944138, 0.023747194558382034, 0.02129397541284561, 0.030242864042520523, -0.08959559351205826, 0.08135181665420532, 0.021293768659234047, -0.20148859918117523, 0.11459939926862717, -0.03189346566796303, -0.08139892667531967, -0.026965856552124023, 0.09429053217172623, -0.0837487280368805, -0.007350565865635872, -0.034845106303691864, 0.025997426360845566, 0.1609819382429123, 0.026628276333212852, -0.012414874508976936, -0.12979178130626678, 0.053646206855773926, -0.15125489234924316, 0.223868265748024, -0.013770811259746552, 0.022791029885411263, 0.05898338928818703, -0.041422829031944275, -0.19374440610408783, -0.004238135647028685, 0.007162882946431637, -0.04089219123125076, 0.011374315246939659, 0.16484102606773376, -0.00496274046599865, 0.19783233106136322, 0.01958032324910164, -0.025192007422447205, -0.005333777517080307, -0.018267372623085976, 0.0011105566518381238, -0.07469144463539124, 0.11589616537094116, -0.10987404733896255, 0.12277135998010635, 0.11425630748271942, -0.023000776767730713, 0.007955324836075306, -0.12881708145141602, 0.02975090965628624, 0.015538466162979603, 0.06706668436527252, 0.06107554957270622, -0.16348686814308167, 0.03425547480583191, -0.0965924859046936, 0.05649358779191971, -0.16336442530155182, -0.14324446022510529, -0.021748170256614685, 0.006131600122898817, -0.048073671758174896, 0.10682682693004608, 0.08085987716913223, 0.020993612706661224, -0.06028050556778908, -0.12839536368846893, 0.034182485193014145, 0.07852669060230255, -0.08294125646352768, -0.0751820057630539 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # cer_model This model is a fine-tuned version of [dmis-lab/biobert-base-cased-v1.1](https://huggingface.co/dmis-lab/biobert-base-cased-v1.1) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.4081 - Precision: 0.9099 - Recall: 0.8471 - F1: 0.8774 - Accuracy: 0.9268 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_ratio: 0.2 - num_epochs: 1 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:| | 0.0179 | 1.0 | 4841 | 0.4081 | 0.9099 | 0.8471 | 0.8774 | 0.9268 | ### Framework versions - Transformers 4.37.0 - Pytorch 2.1.2 - Datasets 2.1.0 - Tokenizers 0.15.1
{"tags": ["generated_from_trainer"], "metrics": ["precision", "recall", "f1", "accuracy"], "base_model": "dmis-lab/biobert-base-cased-v1.1", "model-index": [{"name": "cer_model", "results": []}]}
token-classification
urbija/cer_model
[ "transformers", "tensorboard", "safetensors", "bert", "token-classification", "generated_from_trainer", "base_model:dmis-lab/biobert-base-cased-v1.1", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-10T14:40:29+00:00
[]
[]
TAGS #transformers #tensorboard #safetensors #bert #token-classification #generated_from_trainer #base_model-dmis-lab/biobert-base-cased-v1.1 #autotrain_compatible #endpoints_compatible #region-us
cer\_model ========== This model is a fine-tuned version of dmis-lab/biobert-base-cased-v1.1 on the None dataset. It achieves the following results on the evaluation set: * Loss: 0.4081 * Precision: 0.9099 * Recall: 0.8471 * F1: 0.8774 * Accuracy: 0.9268 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_ratio: 0.2 * num\_epochs: 1 ### Training results ### Framework versions * Transformers 4.37.0 * Pytorch 2.1.2 * Datasets 2.1.0 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.2\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.1.0\n* Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #tensorboard #safetensors #bert #token-classification #generated_from_trainer #base_model-dmis-lab/biobert-base-cased-v1.1 #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.2\n* num\\_epochs: 1", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.1.0\n* Tokenizers 0.15.1" ]
[ 69, 116, 4, 30 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #bert #token-classification #generated_from_trainer #base_model-dmis-lab/biobert-base-cased-v1.1 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_ratio: 0.2\n* num\\_epochs: 1### Training results### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.1.0\n* Tokenizers 0.15.1" ]
[ -0.10136894881725311, 0.08108913153409958, -0.004015001468360424, 0.09952660650014877, 0.12214108556509018, 0.018121706321835518, 0.14201807975769043, 0.1314338743686676, -0.04701366648077965, 0.059128087013959885, 0.13372133672237396, 0.12835732102394104, 0.010994626209139824, 0.13768945634365082, -0.05666935443878174, -0.25425466895103455, 0.00951279979199171, 0.03716493025422096, -0.10437032580375671, 0.10503031313419342, 0.08286596089601517, -0.1337025910615921, 0.09015779942274094, 0.005203794687986374, -0.1466335654258728, 0.012686144560575485, 0.037996385246515274, -0.05330118536949158, 0.1310087889432907, 0.010896825231611729, 0.1417284607887268, 0.03757409751415253, 0.1056368350982666, -0.18223820626735687, 0.008293990045785904, 0.057500410825014114, 0.025425545871257782, 0.09591652452945709, 0.047059837728738785, 0.00620081415399909, 0.08182325214147568, -0.09215544164180756, 0.07868849486112595, 0.016312260180711746, -0.11421653628349304, -0.24068240821361542, -0.06666328758001328, 0.02945205196738243, 0.08101356029510498, 0.05223635956645012, -0.004410531837493181, 0.12697137892246246, -0.04075731709599495, 0.10213635861873627, 0.2557317018508911, -0.28246402740478516, -0.0651453360915184, 0.05441916733980179, 0.054013676941394806, 0.06568998098373413, -0.11014623194932938, -0.009081683121621609, 0.045019105076789856, 0.01932775229215622, 0.146916925907135, -0.030919551849365234, 0.0004011171404272318, 0.005091534927487373, -0.13391470909118652, -0.010440788231790066, 0.1059538796544075, 0.028089310973882675, -0.04769081994891167, -0.055792830884456635, -0.07586660981178284, -0.1741499900817871, -0.04438987746834755, -0.055641964077949524, 0.043533336371183395, -0.05209962651133537, -0.06993963569402695, -0.01722959242761135, -0.08464761078357697, -0.09217223525047302, -0.025722160935401917, 0.20379628241062164, 0.044483911246061325, 0.005245490465313196, -0.01830584928393364, 0.11030212044715881, -0.024542730301618576, -0.15036825835704803, 0.007380199618637562, 0.014346422627568245, -0.021468723192811012, -0.056436166167259216, -0.04099418967962265, -0.04902328550815582, 0.01966029219329357, 0.15227773785591125, -0.07208307832479477, 0.0490146204829216, 0.02293957583606243, 0.010402402840554714, -0.09821155667304993, 0.16400030255317688, -0.03887994587421417, -0.028218504041433334, 0.0036809113807976246, 0.09199708700180054, 0.014258121140301228, -0.0033807391300797462, -0.09775499999523163, 0.01074640266597271, 0.11445234715938568, 0.018078649416565895, -0.0740928202867508, 0.07014383375644684, -0.03779154270887375, -0.02524864859879017, 0.01711372658610344, -0.0967191606760025, 0.0320439375936985, 0.006221150979399681, -0.06522472947835922, -0.052667006850242615, 0.015281982719898224, 0.006487932521849871, 0.01912090554833412, 0.13286516070365906, -0.09878578037023544, 0.012149588204920292, -0.08217527717351913, -0.1289873570203781, -0.009204357862472534, -0.09210240095853806, 0.02092360332608223, -0.08709437400102615, -0.14448757469654083, -0.007110473699867725, 0.053614962846040726, -0.04421475529670715, -0.019676104187965393, -0.03127371147274971, -0.07766293734312057, 0.03228001296520233, -0.003414363833144307, 0.07934171706438065, -0.05657095089554787, 0.09824896603822708, 0.04970509558916092, 0.08489777147769928, -0.012616869993507862, 0.03685648366808891, -0.09592274576425552, 0.048628851771354675, -0.20846085250377655, 0.02092936635017395, -0.07812709361314774, 0.044358640909194946, -0.09185796976089478, -0.08591805398464203, 0.0031354681123048067, -0.004790020175278187, 0.08080442249774933, 0.10504154860973358, -0.1578424572944641, -0.07647643238306046, 0.1617109477519989, -0.0859101414680481, -0.10908520221710205, 0.10889317095279694, -0.061287518590688705, 0.043885525315999985, 0.05001593008637428, 0.18393926322460175, 0.08719022572040558, -0.1304583102464676, 0.006097959354519844, -0.02100979909300804, 0.06436723470687866, 0.006984359119087458, 0.06762982904911041, 0.011335935443639755, 0.01299914438277483, 0.012565708719193935, -0.0408833883702755, 0.029730036854743958, -0.0928492471575737, -0.07815787196159363, -0.035357411950826645, -0.0983768105506897, 0.08261480927467346, 0.05191734433174133, 0.06726067513227463, -0.1063210740685463, -0.09249836206436157, 0.06753767281770706, 0.07668185979127884, -0.06823529303073883, 0.012987622059881687, -0.06723615527153015, 0.06518831849098206, -0.06766185164451599, -0.025028301402926445, -0.15975472331047058, -0.06498687714338303, 0.03227302432060242, -0.01677720993757248, 0.007497136481106281, 0.012784124352037907, 0.0952136218547821, 0.09292569756507874, -0.07213425636291504, -0.02857195772230625, -0.04398259520530701, 0.01519715879112482, -0.12567859888076782, -0.2054041475057602, -0.031374242156744, -0.04119938611984253, 0.08610951155424118, -0.21443067491054535, 0.03789519518613815, 0.0025278162211179733, 0.1110716164112091, 0.06255000829696655, -0.02784557081758976, -0.029040537774562836, 0.04989015683531761, -0.04138919711112976, -0.05909859389066696, 0.04661254957318306, -0.0204667616635561, -0.09147318452596664, -0.05354127660393715, -0.136613667011261, 0.17018617689609528, 0.09951552748680115, -0.04611273109912872, -0.10806863009929657, -0.018287483602762222, -0.055673979222774506, -0.027906836941838264, -0.0473988875746727, 0.0003422381414566189, 0.1370851993560791, -0.0038106145802885294, 0.14255627989768982, -0.07503683120012283, -0.04772685468196869, 0.033006660640239716, -0.029269356280565262, 0.014902891591191292, 0.10625363886356354, 0.07953081279993057, -0.09065704047679901, 0.13434293866157532, 0.16165629029273987, -0.07646326720714569, 0.15130962431430817, -0.03310564160346985, -0.07466145604848862, -0.037502482533454895, -0.00910145789384842, 0.008262486197054386, 0.13157638907432556, -0.10372189432382584, -0.0005183142493478954, 0.016947433352470398, 0.016828281804919243, -0.010091972537338734, -0.19698026776313782, -0.02017679624259472, 0.04655031859874725, -0.05124949663877487, 0.005287796258926392, -0.007445373572409153, 0.009961619973182678, 0.11563409864902496, 0.008863178081810474, -0.07529192417860031, 0.017285892739892006, -0.0014399327337741852, -0.06636248528957367, 0.20032449066638947, -0.06295036524534225, -0.11557367444038391, -0.10810976475477219, -0.051704902201890945, -0.05567912757396698, 0.03098994679749012, 0.043744269758462906, -0.08125041425228119, -0.03189067542552948, -0.07864681631326675, 0.03726206719875336, 0.02508535236120224, 0.05641842633485794, 0.011409495957195759, -0.0038379402831196785, 0.08485143631696701, -0.10199503600597382, -0.011165509931743145, -0.05999192222952843, -0.03538018837571144, 0.029047662392258644, 0.05281652882695198, 0.10103867948055267, 0.1269059032201767, -0.034604255110025406, 0.00848477054387331, -0.02756541408598423, 0.23606754839420319, -0.0742715373635292, -0.0010590038727968931, 0.11740260571241379, -0.028930658474564552, 0.04067362844944, 0.15432652831077576, 0.07011929899454117, -0.08962339162826538, 0.01393887773156166, 0.05640995129942894, -0.03398099169135094, -0.19778169691562653, -0.021449223160743713, -0.023239389061927795, 0.002608232432976365, 0.10873647779226303, 0.020807087421417236, 0.029580647125840187, 0.06946250796318054, 0.025223979726433754, 0.06165117770433426, -0.02691231109201908, 0.07904946058988571, 0.08988089859485626, 0.04303744062781334, 0.13607437908649445, -0.02069489285349846, -0.08405745029449463, 0.029328539967536926, 0.01679014042019844, 0.21317118406295776, 0.008113643154501915, 0.12607651948928833, 0.03567652031779289, 0.13869746029376984, 0.0057021030224859715, 0.07660282403230667, 0.00007515202014474198, -0.052573349326848984, -0.0033400605898350477, -0.05402703955769539, -0.019389040768146515, 0.04376748576760292, -0.07274481654167175, 0.05652623623609543, -0.11047172546386719, 0.02654842846095562, 0.045818548649549484, 0.23755012452602386, 0.052974097430706024, -0.3076181411743164, -0.09956027567386627, 0.015473849140107632, -0.03987690433859825, -0.02690519019961357, 0.026700366288423538, 0.13951066136360168, -0.05614069849252701, 0.040214426815509796, -0.09178013354539871, 0.06583817303180695, -0.04089116305112839, 0.04399022087454796, 0.07142896205186844, 0.09197556972503662, -0.017208369448781013, 0.0503402054309845, -0.250057578086853, 0.2832699120044708, 0.02041459269821644, 0.05448665842413902, -0.05796699970960617, 0.012094874866306782, 0.029802681878209114, 0.0715140625834465, 0.08700773864984512, -0.01921410858631134, -0.0962720587849617, -0.21568326652050018, -0.06861984729766846, 0.010667652823030949, 0.1240849643945694, -0.06202003359794617, 0.12381406873464584, -0.03273197263479233, -0.00946675892919302, 0.06744828075170517, -0.04030732065439224, -0.09134777635335922, -0.06409142911434174, 0.012304356321692467, 0.017980216071009636, -0.03662199154496193, -0.07371760159730911, -0.10494296252727509, -0.09426526725292206, 0.16072572767734528, -0.054023683071136475, -0.035121746361255646, -0.13115185499191284, 0.07174048572778702, 0.09067218005657196, -0.08087436109781265, 0.032169271260499954, 0.004272183403372765, 0.09717585146427155, 0.02761944569647312, -0.061623673886060715, 0.12993896007537842, -0.07429880648851395, -0.20012566447257996, -0.07819360494613647, 0.12625855207443237, 0.027331670746207237, 0.06042991951107979, -0.010501652956008911, 0.024477658793330193, -0.004368823021650314, -0.07378005236387253, 0.05428553745150566, -0.02425120770931244, 0.03701404482126236, 0.0021897219121456146, -0.07506478577852249, 0.029899626970291138, -0.05208015814423561, -0.008711724542081356, 0.14240948855876923, 0.2801716923713684, -0.10536526888608932, 0.02709009498357773, 0.032207097858190536, -0.062288425862789154, -0.194548100233078, 0.041634224355220795, 0.07078928500413895, 0.0025859905872493982, 0.05074457451701164, -0.16623511910438538, 0.11151469498872757, 0.09209097921848297, -0.024097498506307602, 0.11019254475831985, -0.25519251823425293, -0.14967620372772217, 0.11670604348182678, 0.15142670273780823, 0.09719357639551163, -0.15133599936962128, -0.039023108780384064, -0.0014737845631316304, -0.09631381183862686, 0.11184507608413696, -0.08089900761842728, 0.11376616358757019, -0.025100992992520332, 0.07081691920757294, 0.029271846637129784, -0.05898603796958923, 0.11756914854049683, -0.0015090331435203552, 0.10816136747598648, -0.04734479635953903, -0.036680158227682114, 0.03645622357726097, -0.06300392746925354, 0.021628551185131073, -0.06401921808719635, 0.03141448646783829, -0.0560043528676033, -0.03463514521718025, -0.06996505707502365, 0.03797492757439613, -0.035283684730529785, -0.08349341154098511, -0.045279987156391144, 0.03851587325334549, 0.03958015516400337, -0.03033260628581047, 0.16163551807403564, -0.0019673574715852737, 0.1730756163597107, 0.14125072956085205, 0.08650470525026321, -0.04159875586628914, -0.0355105996131897, 0.02167760767042637, -0.01584731787443161, 0.06799071282148361, -0.13649599254131317, 0.05022015795111656, 0.13932012021541595, 0.019821960479021072, 0.12759773433208466, 0.07025328278541565, -0.0281892828643322, -0.00020117955864407122, 0.059881940484046936, -0.17098934948444366, -0.09657220542430878, -0.01111084595322609, -0.06034780293703079, -0.13757793605327606, 0.0792914405465126, 0.1376768797636032, -0.06662575155496597, 0.004082821775227785, -0.009014206007122993, 0.006472043227404356, -0.024711400270462036, 0.21427114307880402, 0.05905263125896454, 0.06344559788703918, -0.07292819768190384, 0.0508054755628109, 0.029629921540617943, -0.07172981649637222, 0.010632465593516827, 0.059088848531246185, -0.07914379984140396, -0.026250408962368965, 0.028430456295609474, 0.18163813650608063, -0.03838905319571495, -0.017982207238674164, -0.16635596752166748, -0.11562884598970413, 0.05762341991066933, 0.19158099591732025, 0.08535843342542648, 0.005939860828220844, -0.03579496592283249, 0.028997180983424187, -0.1187649816274643, 0.13951995968818665, 0.024319319054484367, 0.09725189954042435, -0.16279549896717072, 0.13812203705310822, -0.02511991746723652, 0.018910372629761696, -0.03713059052824974, 0.04526059329509735, -0.12197476625442505, -0.014232072979211807, -0.11952325701713562, -0.02685069479048252, -0.025526907294988632, 0.000032833038858370855, 0.008657270111143589, -0.07753468304872513, -0.05608760192990303, 0.009665845893323421, -0.10663983970880508, -0.016838718205690384, 0.028481842949986458, 0.039111413061618805, -0.1153804287314415, -0.04222874715924263, 0.023948924615979195, -0.07520848512649536, 0.07931111007928848, 0.022327158600091934, 0.028502609580755234, 0.03274545073509216, -0.1233106181025505, 0.03830138221383095, 0.052918024361133575, 0.0034032638650387526, 0.051765140146017075, -0.1113489642739296, -0.016574189066886902, -0.030797259882092476, 0.04624829441308975, 0.015965601429343224, 0.0844118595123291, -0.12977048754692078, 0.003759111510589719, -0.02297157607972622, -0.0645543560385704, -0.056334853172302246, 0.015823274850845337, 0.08109746873378754, -0.0005541989230550826, 0.17874115705490112, -0.08873534202575684, 0.03954008221626282, -0.2126493602991104, -0.0037453314289450645, 0.01020517572760582, -0.10817631334066391, -0.09169460088014603, -0.05332418158650398, 0.06357195228338242, -0.05870601907372475, 0.1299372911453247, 0.00820697471499443, 0.052719876170158386, 0.0401763916015625, -0.0618138462305069, 0.044501833617687225, 0.030731910839676857, 0.1785004884004593, 0.03372562676668167, -0.05204815790057182, 0.032274279743433, 0.036672722548246384, 0.10475431382656097, 0.09795788675546646, 0.2041357457637787, 0.14267702400684357, -0.036462780088186264, 0.0777345821261406, 0.042631056159734726, -0.06545370817184448, -0.1738417148590088, 0.04655710607767105, -0.0413648895919323, 0.09126613289117813, -0.014454870484769344, 0.1947115957736969, 0.0934181958436966, -0.16068746149539948, 0.032814133912324905, -0.03656071051955223, -0.08807380497455597, -0.12566962838172913, -0.038996148854494095, -0.08948130905628204, -0.14733807742595673, 0.003437199629843235, -0.11706896871328354, 0.019515326246619225, 0.06535498052835464, 0.01826697587966919, 0.003745465772226453, 0.20434464514255524, 0.004435091745108366, 0.04474186524748802, 0.048770152032375336, 0.028446996584534645, -0.0286472886800766, -0.07082940638065338, -0.08620774000883102, -0.009523549117147923, -0.017458830028772354, 0.006563524249941111, -0.06929925829172134, -0.05320080369710922, 0.018072057515382767, 0.004625099711120129, -0.10264250636100769, 0.009427866898477077, 0.014120111241936684, 0.04605843126773834, 0.038279689848423004, -0.0005371508304961026, 0.005214807577431202, -0.0244294460862875, 0.2094409465789795, -0.08267496526241302, -0.036879271268844604, -0.10886792093515396, 0.23685063421726227, 0.02410850115120411, 0.02246294915676117, 0.01577623374760151, -0.07684338837862015, 0.009542951360344887, 0.20896290242671967, 0.174658864736557, -0.06238728016614914, 0.00646122545003891, -0.0033840485848486423, -0.012973513454198837, -0.02445179782807827, 0.08375375717878342, 0.09141400456428528, 0.03437180817127228, -0.06902072578668594, -0.055093102157115936, -0.02398078888654709, -0.01627989299595356, -0.05030122399330139, 0.06925832480192184, 0.050344206392765045, 0.022463122382760048, -0.05697103217244148, 0.03566204011440277, -0.02511812001466751, -0.12492368370294571, 0.07489271461963654, -0.20821994543075562, -0.14725863933563232, -0.022311575710773468, 0.08473392575979233, -0.018908169120550156, 0.05020943656563759, -0.02099008485674858, -0.016477715224027634, 0.06778579205274582, -0.018887441605329514, -0.06147947534918785, -0.07947936654090881, 0.08499816805124283, -0.0947079136967659, 0.22535687685012817, -0.03448692336678505, 0.03968942165374756, 0.12278278172016144, 0.02311422862112522, -0.08564898371696472, 0.05512098968029022, 0.06675783544778824, -0.11303254216909409, 0.013230985030531883, 0.14765062928199768, -0.03602329269051552, 0.12404436618089676, 0.04683740809559822, -0.15986360609531403, -0.003468126291409135, -0.05853303149342537, -0.08989402651786804, -0.04611716419458389, -0.028976071625947952, -0.04665500298142433, 0.1413799375295639, 0.21168974041938782, -0.0301953237503767, 0.011090229265391827, -0.05922166630625725, 0.03053952381014824, 0.07562876492738724, 0.04672660306096077, -0.044706858694553375, -0.2532519996166229, 0.029979636892676353, 0.09426532685756683, -0.017940979450941086, -0.260809063911438, -0.1096426323056221, 0.018263744190335274, -0.03398186340928078, -0.08810970187187195, 0.07619868963956833, 0.08875227719545364, 0.06305374205112457, -0.050690729171037674, -0.10355407744646072, -0.06077003851532936, 0.1528947949409485, -0.15505872666835785, -0.0829538032412529 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-finetuned-emotion This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the emotion dataset. It achieves the following results on the evaluation set: - Loss: 0.2139 - Accuracy: 0.9235 - F1: 0.9235 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 64 - eval_batch_size: 64 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | 0.7853 | 1.0 | 250 | 0.3142 | 0.908 | 0.9068 | | 0.2433 | 2.0 | 500 | 0.2139 | 0.9235 | 0.9235 | ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.1
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["emotion"], "metrics": ["accuracy", "f1"], "base_model": "distilbert-base-uncased", "model-index": [{"name": "distilbert-base-uncased-finetuned-emotion", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "emotion", "type": "emotion", "config": "split", "split": "validation", "args": "split"}, "metrics": [{"type": "accuracy", "value": 0.9235, "name": "Accuracy"}, {"type": "f1", "value": 0.9235223901734632, "name": "F1"}]}]}]}
text-classification
umeboshi827/distilbert-base-uncased-finetuned-emotion
[ "transformers", "tensorboard", "safetensors", "distilbert", "text-classification", "generated_from_trainer", "dataset:emotion", "base_model:distilbert-base-uncased", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-10T14:41:03+00:00
[]
[]
TAGS #transformers #tensorboard #safetensors #distilbert #text-classification #generated_from_trainer #dataset-emotion #base_model-distilbert-base-uncased #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
distilbert-base-uncased-finetuned-emotion ========================================= This model is a fine-tuned version of distilbert-base-uncased on the emotion dataset. It achieves the following results on the evaluation set: * Loss: 0.2139 * Accuracy: 0.9235 * F1: 0.9235 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 64 * eval\_batch\_size: 64 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 2 ### Training results ### Framework versions * Transformers 4.35.2 * Pytorch 2.1.0+cu121 * Datasets 2.17.0 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2", "### Training results", "### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #tensorboard #safetensors #distilbert #text-classification #generated_from_trainer #dataset-emotion #base_model-distilbert-base-uncased #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2", "### Training results", "### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ 82, 98, 4, 33 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #distilbert #text-classification #generated_from_trainer #dataset-emotion #base_model-distilbert-base-uncased #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 64\n* eval\\_batch\\_size: 64\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ -0.12298925966024399, 0.16415931284427643, -0.0022249515168368816, 0.12838715314865112, 0.12877148389816284, 0.01904618926346302, 0.1514195203781128, 0.12019442021846771, -0.03650584816932678, 0.046193916350603104, 0.12899209558963776, 0.13060440123081207, 0.020094528794288635, 0.13745972514152527, -0.08555696159601212, -0.20673809945583344, 0.021901700645685196, 0.034984759986400604, -0.019554216414690018, 0.1276279240846634, 0.10021701455116272, -0.11031892150640488, 0.10418032854795456, -0.0035885514225810766, -0.13512423634529114, -0.007379281334578991, 0.016138669103384018, -0.038977932184934616, 0.12101870775222778, 0.019757069647312164, 0.08555757254362106, 0.03709016367793083, 0.061780042946338654, -0.2063295543193817, 0.01633494906127453, 0.037308480590581894, -0.006699043791741133, 0.08305735141038895, 0.033081576228141785, -0.027233779430389404, 0.06545687466859818, -0.09334614872932434, 0.04899849742650986, 0.025416752323508263, -0.12426608800888062, -0.2230583131313324, -0.08652891963720322, 0.05607512220740318, 0.06954199075698853, 0.09302328526973724, -0.02298552542924881, 0.12882556021213531, -0.03278733789920807, 0.09309118241071701, 0.1929847002029419, -0.2785753309726715, -0.06253271549940109, 0.019933240488171577, 0.024857191368937492, 0.07781456410884857, -0.11607067286968231, -0.02765176258981228, 0.05328065901994705, 0.025419652462005615, 0.1408209651708603, -0.03000331111252308, -0.0000100845136330463, -0.008367571979761124, -0.122127965092659, -0.0467156320810318, 0.19661131501197815, 0.08830983191728592, -0.05298612639307976, -0.08292418718338013, -0.062137067317962646, -0.12344267219305038, -0.02369128353893757, -0.015646615996956825, 0.048551563173532486, -0.0006837965920567513, -0.07165957242250443, -0.018645696341991425, -0.10995084792375565, -0.039818160235881805, -0.014501640573143959, 0.13194607198238373, 0.011023708619177341, 0.0013879250036552548, 0.006495350040495396, 0.09365515410900116, -0.005687214434146881, -0.15282344818115234, 0.011868472211062908, 0.008473686873912811, 0.02234680764377117, -0.03086315095424652, -0.06160522252321243, -0.08252810686826706, 0.007355652283877134, 0.12234635651111603, -0.06391282379627228, 0.0381796732544899, 0.02186567336320877, 0.02776717022061348, -0.0757814571261406, 0.18540902435779572, -0.03730515390634537, -0.04703554883599281, 0.018810899928212166, 0.12801603972911835, 0.06705563515424728, -0.01113202329725027, -0.11687757074832916, 0.036215994507074356, 0.12639904022216797, 0.008259239606559277, -0.06383469700813293, 0.07295916974544525, -0.09291710704565048, -0.03045615740120411, 0.040017351508140564, -0.08345825970172882, 0.010006575845181942, 0.010459926910698414, -0.051640018820762634, -0.05533228814601898, 0.02033049426972866, 0.023422755300998688, 0.008880945853888988, 0.05069858580827713, -0.08697323501110077, -0.004888592287898064, -0.06064574047923088, -0.09836641699075699, 0.013786204159259796, -0.07303544878959656, 0.0294688418507576, -0.10141290724277496, -0.22318731248378754, -0.023288331925868988, 0.06568637490272522, -0.015711693093180656, -0.05751540884375572, -0.0765494853258133, -0.05566942319273949, 0.01963070221245289, -0.0009356812224723399, 0.02586907520890236, -0.06492088735103607, 0.08844626694917679, 0.043852370232343674, 0.06694947183132172, -0.04401993751525879, 0.03836948052048683, -0.12762954831123352, 0.04506474360823631, -0.12825657427310944, 0.064083531498909, -0.03988830745220184, 0.08981011807918549, -0.07992232590913773, -0.07603347301483154, 0.015640147030353546, -0.02602706104516983, 0.05740893632173538, 0.1261247992515564, -0.16774190962314606, -0.07420198619365692, 0.16842208802700043, -0.08345673978328705, -0.15481406450271606, 0.13748565316200256, -0.05880170688033104, 0.07106537371873856, 0.07528718560934067, 0.20262359082698822, 0.06439574062824249, -0.048542819917201996, -0.022134937345981598, -0.005324368830770254, 0.07184375822544098, -0.00817684456706047, 0.09434540569782257, 0.008535358123481274, -0.004509931895881891, 0.022837214171886444, -0.05818382650613785, 0.060154400765895844, -0.06362628936767578, -0.10318101197481155, -0.03899693489074707, -0.1149534285068512, 0.07968122512102127, 0.06385963410139084, 0.05263492837548256, -0.10782752186059952, -0.08684656769037247, 0.011198734864592552, 0.0920170322060585, -0.08252817392349243, 0.01560035441070795, -0.0750403180718422, 0.0779593288898468, -0.06215186044573784, -0.012691549956798553, -0.14994220435619354, 0.0015186409000307322, 0.025656571611762047, 0.008467240259051323, -0.003673878498375416, -0.010716867633163929, 0.07532812654972076, 0.057596828788518906, -0.07332060486078262, -0.06730838119983673, -0.036077093333005905, 0.009738643653690815, -0.09620614349842072, -0.20298953354358673, -0.013542743399739265, -0.04094722494482994, 0.2085779309272766, -0.22458098828792572, 0.055817391723394394, 0.001140425680205226, 0.06879791617393494, 0.03843928128480911, -0.03587650880217552, -0.0029504087287932634, 0.03329670801758766, -0.04915175959467888, -0.07146980613470078, 0.07085344940423965, 0.02826550044119358, -0.1311282217502594, -0.0149794090539217, -0.14575673639774323, 0.16261208057403564, 0.11418330669403076, -0.03650867938995361, -0.048777587711811066, -0.0063068438321352005, -0.04165167361497879, -0.020295223221182823, -0.018865853548049927, 0.009229526855051517, 0.1351911425590515, 0.007918436080217361, 0.1456819772720337, -0.08592311292886734, -0.021741749718785286, 0.017022565007209778, -0.04692652076482773, -0.013722056522965431, 0.11608871072530746, 0.013309158384799957, -0.14549824595451355, 0.14500054717063904, 0.1914234459400177, -0.06624388694763184, 0.14039725065231323, -0.042862098664045334, -0.04359167069196701, -0.049233049154281616, -0.00044229652849026024, 0.008725951425731182, 0.10347870737314224, -0.11197403818368912, 0.002216064603999257, 0.01330116018652916, -0.00366339017637074, -0.00963506568223238, -0.19666317105293274, -0.038123928010463715, 0.06050406023859978, -0.04967494308948517, 0.004861168097704649, -0.009077387861907482, -0.02050962671637535, 0.08195244520902634, 0.005004735663533211, -0.06806423515081406, 0.05337528884410858, -0.008175288327038288, -0.08379372954368591, 0.19792059063911438, -0.08355582505464554, -0.1820179522037506, -0.1360347718000412, -0.04509621486067772, -0.08249577134847641, 0.03641597554087639, 0.06329063326120377, -0.07827824354171753, -0.02807299606502056, -0.11779822409152985, -0.022146549075841904, 0.026902813464403152, 0.01316885743290186, 0.046345897018909454, -0.02449585497379303, 0.09087736159563065, -0.09302917122840881, -0.015289775095880032, -0.015670262277126312, -0.02726726606488228, 0.043986979871988297, 0.0003908520156983286, 0.11644674837589264, 0.1344602108001709, -0.0024111997336149216, 0.001982764108106494, -0.02674761414527893, 0.24015909433364868, -0.060976624488830566, -0.021268535405397415, 0.14057914912700653, -0.025824066251516342, 0.0664115697145462, 0.14003711938858032, 0.04754038155078888, -0.10216362029314041, 0.01825735531747341, 0.02516135387122631, -0.022555982694029808, -0.19683779776096344, -0.023665999993681908, -0.03820670023560524, 0.013110728934407234, 0.0923033356666565, 0.02500477060675621, 0.05578959360718727, 0.07980433851480484, 0.011307960376143456, 0.046082641929388046, -0.01593147963285446, 0.07934863120317459, 0.11102430522441864, 0.03545748442411423, 0.10672835260629654, -0.02969278022646904, -0.036512281745672226, 0.05040682479739189, -0.005518865305930376, 0.17186596989631653, 0.0012750305468216538, 0.19665370881557465, 0.0376354418694973, 0.16429176926612854, -0.03006909415125847, 0.057630978524684906, -0.009824772365391254, -0.02506519854068756, -0.029358426108956337, -0.042727939784526825, -0.07352055609226227, 0.047611966729164124, -0.06342733651399612, 0.09695445001125336, -0.11881578713655472, 0.016130482777953148, 0.06686265766620636, 0.27346566319465637, 0.05122973397374153, -0.3381480276584625, -0.118779256939888, 0.03163718804717064, -0.012638452462852001, -0.024628782644867897, 0.006265256088227034, 0.11430396884679794, -0.0633852407336235, 0.053050447255373, -0.07694634050130844, 0.0762980729341507, -0.06483963876962662, 0.05347886681556702, 0.017030466347932816, 0.05816536024212837, -0.003607850056141615, 0.07206951081752777, -0.2507160007953644, 0.237730011343956, 0.011494971811771393, 0.06829703599214554, -0.05151974409818649, -0.004263938404619694, 0.06421706080436707, 0.0902511328458786, 0.08569177985191345, -0.0013113958993926644, 0.00467272661626339, -0.17602375149726868, -0.06630519032478333, 0.022091830149292946, 0.0539868026971817, -0.06566138565540314, 0.09515029937028885, -0.03607471287250519, 0.005177485756576061, 0.06704428791999817, 0.042426832020282745, -0.07040222734212875, -0.10223818570375443, 0.002481512725353241, 0.05932764708995819, 0.0026643802411854267, -0.08312251418828964, -0.10207563638687134, -0.09500055015087128, 0.15390591323375702, -0.012838419526815414, -0.04031655564904213, -0.1032203808426857, 0.05126528441905975, 0.04013681784272194, -0.08564738184213638, 0.019136466085910797, -0.003202908206731081, 0.11475201696157455, 0.01415063627064228, -0.05189846456050873, 0.10044047981500626, -0.06313786655664444, -0.1765647828578949, -0.05559659004211426, 0.10916933417320251, 0.02679434046149254, 0.04905441030859947, 0.0072315763682127, 0.006358620710670948, -0.04452360421419144, -0.06758596748113632, 0.04355176165699959, 0.02181336283683777, 0.04135998710989952, 0.01441817544400692, -0.01724773831665516, -0.004459582734853029, -0.07943835109472275, -0.029633482918143272, 0.16276660561561584, 0.2990800440311432, -0.0724276676774025, 0.008852411061525345, 0.053898707032203674, -0.053623735904693604, -0.17211347818374634, 0.02427157387137413, 0.03177283704280853, 0.015426634810864925, 0.0598289929330349, -0.14455188810825348, 0.0749862790107727, 0.06510473042726517, -0.0287938192486763, 0.0798645168542862, -0.2513330578804016, -0.12475714832544327, 0.12642055749893188, 0.14921393990516663, 0.1332807093858719, -0.15596584975719452, -0.038069210946559906, -0.04890201613306999, -0.10739458352327347, 0.10282620787620544, -0.10546713322401047, 0.10821212828159332, -0.006642368156462908, 0.07144784927368164, 0.012308468110859394, -0.037015244364738464, 0.1497841775417328, -0.0053465948440134525, 0.10033758729696274, -0.05874332785606384, -0.012331431731581688, 0.07007947564125061, -0.07364233583211899, 0.03307785093784332, -0.140011727809906, 0.04514949768781662, -0.10610543936491013, -0.03220716491341591, -0.06697490066289902, 0.02192913554608822, -0.03276041895151138, -0.06069423258304596, -0.03603912517428398, 0.04129990562796593, 0.09074567258358002, 0.004461248405277729, 0.13149239122867584, 0.012619505636394024, 0.11043127626180649, 0.15007495880126953, 0.09350668638944626, -0.045194126665592194, -0.03603368252515793, -0.031068192794919014, -0.0329701267182827, 0.051759686321020126, -0.15761122107505798, 0.038663070648908615, 0.10641254484653473, 0.008402816019952297, 0.17831099033355713, 0.05722370743751526, -0.0326821506023407, 0.015341658145189285, 0.061104100197553635, -0.1644946038722992, -0.12632444500923157, -0.03550735488533974, -0.023713625967502594, -0.16105681657791138, 0.02984657511115074, 0.11732334643602371, -0.05834845080971718, -0.0014169812202453613, -0.015113011002540588, 0.02170756831765175, -0.021074876189231873, 0.14048346877098083, 0.04846009239554405, 0.03503165766596794, -0.09008875489234924, 0.09292999655008316, 0.039195068180561066, -0.0952715054154396, 0.0248859915882349, 0.012673599645495415, -0.0900382325053215, -0.04714033380150795, 0.03031144291162491, 0.19745920598506927, -0.040961854159832, -0.051432423293590546, -0.15959081053733826, -0.10266458988189697, 0.051544420421123505, 0.1154305636882782, 0.09952162951231003, 0.023818695917725563, -0.04972377419471741, -0.0019818672444671392, -0.09349818527698517, 0.11611457169055939, 0.06941138207912445, 0.061568208038806915, -0.16531093418598175, 0.0715322345495224, -0.01668120175600052, 0.014330887235701084, -0.016147714108228683, 0.01841355860233307, -0.07997079938650131, -0.019082441926002502, -0.13620004057884216, 0.0056248800829052925, -0.04218827933073044, 0.017805742099881172, -0.0014917199732735753, -0.05678331479430199, -0.04375004023313522, 0.010821771807968616, -0.09677516669034958, -0.03369671106338501, 0.04078563302755356, 0.0690656453371048, -0.11939346790313721, -0.051640816032886505, 0.029735703021287918, -0.07973416894674301, 0.08691716939210892, 0.03120613284409046, 0.009364795871078968, 0.03232576325535774, -0.15181834995746613, 0.04097861424088478, 0.061356496065855026, 0.0010428341338410974, 0.018393993377685547, -0.10278753936290741, -0.026722000911831856, -0.0025697823148220778, 0.007415526546537876, 0.016234593465924263, 0.10535606741905212, -0.11115217953920364, 0.007499100640416145, 0.011769969016313553, -0.037147048860788345, -0.06605912744998932, 0.037295207381248474, 0.07689788937568665, 0.021435894072055817, 0.22531521320343018, -0.08321985602378845, 0.01358514092862606, -0.20898322761058807, 0.008152627386152744, 0.004840954206883907, -0.11456567794084549, -0.1384253352880478, -0.06088176742196083, 0.03864532336592674, -0.05609184131026268, 0.1072925254702568, -0.005791793577373028, 0.02195131406188011, 0.019817862659692764, 0.0024811180774122477, 0.05772358551621437, 0.00891666579991579, 0.2178145945072174, 0.016372209414839745, -0.04919351637363434, 0.0604318343102932, 0.027111297473311424, 0.11198686808347702, 0.11162023991346359, 0.1395191252231598, 0.1593293696641922, -0.004752764478325844, 0.10206945240497589, 0.011167536489665508, 0.003676127642393112, -0.14276158809661865, 0.04454519972205162, -0.03397553041577339, 0.10183025151491165, 0.004331657197326422, 0.23346753418445587, 0.08796288818120956, -0.16622917354106903, 0.03672078251838684, -0.04876039922237396, -0.07527824491262436, -0.07965926080942154, -0.10067258030176163, -0.09405376762151718, -0.13725221157073975, -0.005322910379618406, -0.12106827646493912, 0.0037962363567203283, 0.07040152698755264, -0.00980472657829523, -0.04000251367688179, 0.14826920628547668, -0.004319100175052881, 0.01474982313811779, 0.08450791984796524, -0.01017678715288639, -0.07189807295799255, -0.06506330519914627, -0.08639872819185257, 0.020230503752827644, -0.0006633383454754949, 0.04267972707748413, -0.04681497439742088, -0.024334469810128212, 0.030356328934431076, -0.016857890412211418, -0.12031431496143341, 0.010171568021178246, 0.02395612746477127, 0.04880282282829285, 0.044744398444890976, 0.015952477231621742, 0.00639567943289876, 0.01082262210547924, 0.2127608209848404, -0.06947115808725357, -0.01553473062813282, -0.1102161779999733, 0.16648617386817932, 0.0038185124285519123, -0.013057523407042027, 0.02416670322418213, -0.09783343225717545, 0.042867351323366165, 0.1887388378381729, 0.16507670283317566, -0.09282473474740982, 0.004849682096391916, -0.03368780389428139, -0.0017054337076842785, -0.04403488710522652, 0.06038481742143631, 0.09466515481472015, -0.05141542851924896, -0.08203145116567612, 0.002195686334744096, -0.04958672448992729, -0.023554332554340363, -0.023252420127391815, 0.04895400255918503, 0.02790220081806183, 0.01331760548055172, -0.05317055433988571, 0.055212270468473434, -0.04259690269827843, -0.08908868581056595, 0.05131212994456291, -0.1913837343454361, -0.13959868252277374, -0.04022086411714554, 0.07758738100528717, 0.016668090596795082, 0.053110647946596146, -0.011013668961822987, 0.0170731358230114, 0.08375632762908936, -0.02727385051548481, -0.07615333795547485, -0.057335834950208664, 0.07826893031597137, -0.10985586792230606, 0.2154049277305603, -0.046016622334718704, 0.01907510496675968, 0.12245169281959534, 0.03985793516039848, -0.0819580927491188, 0.08338607102632523, 0.046950891613960266, -0.02591324783861637, 0.033736422657966614, 0.10738630592823029, -0.03027603216469288, 0.12614227831363678, 0.05635765939950943, -0.12484896183013916, 0.002519885776564479, -0.021606534719467163, -0.07779785990715027, -0.05144015699625015, -0.02547837793827057, -0.05639272183179855, 0.13560938835144043, 0.18014462292194366, -0.05217359960079193, -0.003382803173735738, -0.04032931104302406, 0.01732526533305645, 0.07948071509599686, 0.016613582149147987, -0.03237352892756462, -0.20988033711910248, 0.01582900993525982, 0.07813900709152222, 0.009509092196822166, -0.3007434606552124, -0.10395227372646332, -0.01963624358177185, -0.04528871551156044, -0.06816833466291428, 0.09115426242351532, 0.08285064995288849, 0.048586610704660416, -0.051645874977111816, -0.054796792566776276, -0.07190260291099548, 0.16186556220054626, -0.11737532913684845, -0.089300237596035 ]
null
null
peft
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl) <details><summary>See axolotl config</summary> axolotl version: `0.4.0` ```yaml base_model: Crystalcareai/CrystalMistral-26b model_type: AutoModelForCausalLM tokenizer_type: LlamaTokenizer trust_remote_code: true load_in_8bit: false load_in_4bit: true strict: false datasets: - path: Crystalcareai/synthetic_reasoning_natural_Alpaca_Combined type: alpaca dataset_prepared_path: last_run_prepared val_set_size: 0.0 output_dir: ./qlora-out ## You can optionally freeze the entire model and unfreeze a subset of parameters unfrozen_parameters: # - lm_head.* # - model.embed_tokens.* # - model.layers.2[0-9]+.block_sparse_moe.gate.* # - model.layers.2[0-9]+.block_sparse_moe.experts.* # - model.layers.3[0-9]+.block_sparse_moe.gate.* # - model.layers.3[0-9]+.block_sparse_moe.experts.* model_config: output_router_logits: true adapter: qlora lora_model_dir: sequence_len: 4096 sample_packing: true pad_to_sequence_len: true lora_r: 32 lora_alpha: 16 lora_dropout: 0.05 lora_target_linear: true lora_fan_in_fan_out: #lora_target_modules: # - gate # - q_proj # - k_proj # - v_proj # - o_proj # - w1 # - w2 # - w3 wandb_project: wandb_entity: wandb_watch: wandb_name: wandb_log_model: gradient_accumulation_steps: 16 micro_batch_size: 1 num_epochs: 2.2 optimizer: adamw_bnb_8bit lr_scheduler: cosine learning_rate: 0.0005 train_on_inputs: false group_by_length: false bf16: auto fp16: tf32: false gradient_checkpointing: true early_stopping_patience: resume_from_checkpoint: local_rank: logging_steps: 1 xformers_attention: flash_attention: true loss_watchdog_threshold: 5.0 loss_watchdog_patience: 3 warmup_steps: 10 evals_per_epoch: 4 eval_table_size: eval_table_max_new_tokens: 128 saves_per_epoch: 1 debug: deepspeed: weight_decay: 0.0 fsdp: fsdp_config: special_tokens: ``` </details><br> # qlora-out This model is a fine-tuned version of [Crystalcareai/CrystalMistral-26b](https://huggingface.co/Crystalcareai/CrystalMistral-26b) on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 1 - eval_batch_size: 1 - seed: 42 - distributed_type: multi-GPU - num_devices: 4 - gradient_accumulation_steps: 16 - total_train_batch_size: 64 - total_eval_batch_size: 4 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 10 - num_epochs: 2.2 ### Training results ### Framework versions - PEFT 0.8.2 - Transformers 4.38.0.dev0 - Pytorch 2.1.2+cu118 - Datasets 2.17.0 - Tokenizers 0.15.0
{"license": "apache-2.0", "library_name": "peft", "tags": ["generated_from_trainer"], "base_model": "Crystalcareai/CrystalMistral-26b", "model-index": [{"name": "qlora-out", "results": []}]}
null
Crystalcareai/Crystalmistral-26B-Lora
[ "peft", "tensorboard", "safetensors", "mixtral", "generated_from_trainer", "base_model:Crystalcareai/CrystalMistral-26b", "license:apache-2.0", "4-bit", "region:us" ]
2024-02-10T14:47:36+00:00
[]
[]
TAGS #peft #tensorboard #safetensors #mixtral #generated_from_trainer #base_model-Crystalcareai/CrystalMistral-26b #license-apache-2.0 #4-bit #region-us
<img src="URL alt="Built with Axolotl" width="200" height="32"/> <details><summary>See axolotl config</summary> axolotl version: '0.4.0' </details><br> # qlora-out This model is a fine-tuned version of Crystalcareai/CrystalMistral-26b on the None dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.0005 - train_batch_size: 1 - eval_batch_size: 1 - seed: 42 - distributed_type: multi-GPU - num_devices: 4 - gradient_accumulation_steps: 16 - total_train_batch_size: 64 - total_eval_batch_size: 4 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - lr_scheduler_warmup_steps: 10 - num_epochs: 2.2 ### Training results ### Framework versions - PEFT 0.8.2 - Transformers 4.38.0.dev0 - Pytorch 2.1.2+cu118 - Datasets 2.17.0 - Tokenizers 0.15.0
[ "# qlora-out\n\nThis model is a fine-tuned version of Crystalcareai/CrystalMistral-26b on the None dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0005\n- train_batch_size: 1\n- eval_batch_size: 1\n- seed: 42\n- distributed_type: multi-GPU\n- num_devices: 4\n- gradient_accumulation_steps: 16\n- total_train_batch_size: 64\n- total_eval_batch_size: 4\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_steps: 10\n- num_epochs: 2.2", "### Training results", "### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.2+cu118\n- Datasets 2.17.0\n- Tokenizers 0.15.0" ]
[ "TAGS\n#peft #tensorboard #safetensors #mixtral #generated_from_trainer #base_model-Crystalcareai/CrystalMistral-26b #license-apache-2.0 #4-bit #region-us \n", "# qlora-out\n\nThis model is a fine-tuned version of Crystalcareai/CrystalMistral-26b on the None dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0005\n- train_batch_size: 1\n- eval_batch_size: 1\n- seed: 42\n- distributed_type: multi-GPU\n- num_devices: 4\n- gradient_accumulation_steps: 16\n- total_train_batch_size: 64\n- total_eval_batch_size: 4\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_steps: 10\n- num_epochs: 2.2", "### Training results", "### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.2+cu118\n- Datasets 2.17.0\n- Tokenizers 0.15.0" ]
[ 57, 34, 6, 12, 8, 3, 157, 4, 44 ]
[ "passage: TAGS\n#peft #tensorboard #safetensors #mixtral #generated_from_trainer #base_model-Crystalcareai/CrystalMistral-26b #license-apache-2.0 #4-bit #region-us \n# qlora-out\n\nThis model is a fine-tuned version of Crystalcareai/CrystalMistral-26b on the None dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 0.0005\n- train_batch_size: 1\n- eval_batch_size: 1\n- seed: 42\n- distributed_type: multi-GPU\n- num_devices: 4\n- gradient_accumulation_steps: 16\n- total_train_batch_size: 64\n- total_eval_batch_size: 4\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: cosine\n- lr_scheduler_warmup_steps: 10\n- num_epochs: 2.2### Training results### Framework versions\n\n- PEFT 0.8.2\n- Transformers 4.38.0.dev0\n- Pytorch 2.1.2+cu118\n- Datasets 2.17.0\n- Tokenizers 0.15.0" ]
[ -0.10255884379148483, 0.11905857175588608, -0.0030905415769666433, 0.08076983690261841, 0.11780809611082077, 0.03208485245704651, 0.08743969351053238, 0.14113302528858185, -0.07277835160493851, 0.12981557846069336, 0.0800435021519661, 0.029152020812034607, 0.060457196086645126, 0.16139018535614014, 0.0009967117803171277, -0.24301381409168243, -0.011618402786552906, -0.041467197239398956, -0.08718439936637878, 0.08506881445646286, 0.10769793391227722, -0.09481161832809448, 0.06899657845497131, 0.00027728243730962276, -0.11218541115522385, -0.011243081651628017, -0.0520089715719223, -0.038959018886089325, 0.09613800048828125, 0.010789317078888416, 0.06350569427013397, -0.015205501578748226, 0.11642614752054214, -0.2603263854980469, -0.0036747565027326345, 0.09841594099998474, 0.036859769374132156, 0.0855732262134552, 0.06709454953670502, 0.04272671043872833, 0.11201126873493195, -0.18012994527816772, 0.09613429754972458, 0.04303806647658348, -0.07064054161310196, -0.14966432750225067, -0.0948927104473114, 0.08970007300376892, 0.1145993322134018, 0.08768142014741898, 0.006100763101130724, 0.13725389540195465, -0.07044696807861328, 0.0686284750699997, 0.18352119624614716, -0.2669118642807007, -0.08902120590209961, 0.044294003397226334, 0.060939591377973557, 0.0871727392077446, -0.12310135364532471, -0.0035699689760804176, 0.030304402112960815, 0.03292885050177574, 0.09266532957553864, 0.009228923358023167, 0.021933507174253464, -0.003682459471747279, -0.12029074132442474, -0.011677289381623268, 0.11493884027004242, 0.06643915176391602, -0.029476223513484, -0.15653637051582336, -0.052412379533052444, -0.16237282752990723, -0.00875833723694086, -0.024434581398963928, 0.031488366425037384, -0.044113945215940475, -0.024921758100390434, -0.03410244733095169, -0.05839892476797104, -0.06771547347307205, 0.040898386389017105, 0.10022014379501343, 0.04564704746007919, 0.0020298389717936516, 0.007895458489656448, 0.10368786752223969, -0.0038935900665819645, -0.10647068917751312, -0.03190360963344574, -0.009829591028392315, -0.11921051144599915, -0.03914427384734154, -0.03310014680027962, 0.04706580564379692, -0.0011388834100216627, 0.15317873656749725, -0.051706135272979736, 0.09971719235181808, 0.08464610576629639, -0.0012854645028710365, -0.03829394653439522, 0.13399307429790497, -0.031804729253053665, -0.059448860585689545, -0.02705490216612816, 0.1003492996096611, 0.02800631709396839, -0.012722441926598549, -0.045671913772821426, -0.008447465486824512, 0.05360160022974014, 0.05562246963381767, -0.03440412878990173, -0.006747727748006582, -0.043133191764354706, -0.0314156673848629, 0.07410355657339096, -0.10518467426300049, 0.06106426194310188, 0.01331652607768774, -0.042559511959552765, -0.013552157208323479, -0.014048547483980656, 0.0023133486974984407, -0.03379763290286064, 0.09707794338464737, -0.07919413596391678, -0.03219529613852501, -0.06509121507406235, -0.03408908098936081, 0.010841109789907932, -0.06754438579082489, -0.01510077714920044, -0.06704480201005936, -0.11605481058359146, -0.04844445362687111, 0.04890427365899086, -0.0875396579504013, -0.080730140209198, -0.015639115124940872, -0.06176503002643585, 0.054816149175167084, 0.009750974364578724, 0.13813906908035278, -0.03399330750107765, 0.06842852383852005, -0.018551291897892952, 0.018985506147146225, 0.04514265060424805, 0.028463903814554214, -0.04417106509208679, 0.053570378571748734, -0.08545131981372833, 0.06597466766834259, -0.08243897557258606, 0.013106267899274826, -0.13892193138599396, -0.09534484148025513, -0.02394157461822033, -0.03209252655506134, 0.058627404272556305, 0.10287369042634964, -0.16493020951747894, -0.02457987517118454, 0.1525731235742569, -0.05736294388771057, -0.09245599806308746, 0.09362097084522247, -0.014100567437708378, -0.04157331585884094, 0.028562625870108604, 0.12559349834918976, 0.12169434875249863, -0.15228238701820374, -0.045661307871341705, 0.025217512622475624, 0.08768104761838913, 0.015541031956672668, 0.10168813914060593, -0.03643671050667763, 0.06546859443187714, 0.004356998950242996, -0.07260926812887192, -0.00747933192178607, -0.06110250949859619, -0.07208037376403809, -0.056104984134435654, -0.08194301277399063, 0.00722583569586277, 0.015558942221105099, 0.01977626606822014, -0.06369055807590485, -0.1332486867904663, 0.02484053745865822, 0.1403616964817047, -0.04413459822535515, 0.004760391544550657, -0.061724137514829636, 0.03691177815198898, -0.0077899121679365635, -0.021116498857736588, -0.15146201848983765, -0.1128237396478653, 0.05739603936672211, -0.11248704046010971, -0.012482423335313797, -0.021414702758193016, 0.05509741231799126, 0.08299560844898224, -0.05926879122853279, -0.02364911139011383, -0.09618166089057922, -0.012690519914031029, -0.09270790219306946, -0.17582906782627106, -0.06320735812187195, -0.031158296391367912, 0.20117999613285065, -0.224245086312294, 0.006985079031437635, 0.025730794295668602, 0.1483771800994873, 0.033171989023685455, -0.0794237032532692, -0.018282564356923103, 0.04206976294517517, 0.007817240431904793, -0.11573898047208786, 0.03724436089396477, 0.006220616865903139, -0.09319286793470383, -0.06178661808371544, -0.14605045318603516, 0.02079259417951107, 0.06995397806167603, 0.12702442705631256, -0.08690100163221359, -0.08909580111503601, -0.07314922660589218, -0.045124925673007965, -0.05258714407682419, -0.0036669825203716755, 0.17013660073280334, 0.04279567301273346, 0.10777599364519119, -0.05187363550066948, -0.08075893670320511, 0.02215874381363392, 0.041400808840990067, -0.02993335761129856, 0.09480655193328857, 0.07726509869098663, -0.10887963324785233, 0.062300942838191986, 0.09226372838020325, -0.026117591187357903, 0.1266736537218094, -0.05752303823828697, -0.09168560057878494, -0.02994976006448269, 0.03315296396613121, 0.01686101034283638, 0.14013607800006866, -0.03715309500694275, 0.02150341309607029, 0.023327026516199112, 0.014257756061851978, 0.020723823457956314, -0.1532258838415146, -0.008226806297898293, 0.029154615476727486, -0.013218031264841557, -0.0211437176913023, -0.015176435001194477, -0.0057964688166975975, 0.06019490584731102, 0.03944600000977516, 0.02087131328880787, 0.0034762704744935036, -0.01083663385361433, -0.08566664159297943, 0.1690182238817215, -0.1191238984465599, -0.10333159565925598, -0.13821160793304443, 0.07021858543157578, -0.032948944717645645, -0.04442545026540756, 0.002352448645979166, -0.08639344573020935, -0.04359617084264755, -0.09576644748449326, -0.025787075981497765, -0.04970850050449371, 0.003135515609756112, 0.05643903464078903, 0.020548688247799873, 0.09021052718162537, -0.1148359552025795, 0.019446365535259247, 0.0028425364289432764, -0.05521664395928383, -0.003486780682578683, 0.051640551537275314, 0.07819422334432602, 0.11916196346282959, 0.014147928915917873, 0.022095371037721634, -0.025482336059212685, 0.2392195761203766, -0.10969559848308563, 0.024419842287898064, 0.10235707461833954, 0.04649576172232628, 0.06353909522294998, 0.10457427054643631, 0.04718557745218277, -0.06923991441726685, 0.02146226353943348, 0.08212125301361084, -0.015317429788410664, -0.22303366661071777, -0.05072837695479393, -0.03732877969741821, -0.08909265697002411, 0.1220444068312645, 0.07555129379034042, -0.03341110423207283, 0.0402514711022377, -0.021973857656121254, 0.03609335795044899, -0.02036949433386326, 0.085310198366642, 0.06603232026100159, 0.04453802481293678, 0.08134482055902481, -0.02107411064207554, 0.006405731197446585, 0.06627658754587173, 0.02045942284166813, 0.22964011132717133, -0.034628286957740784, 0.11824911832809448, 0.012768923304975033, 0.13640843331813812, -0.029960403218865395, 0.026336925104260445, 0.02972964011132717, -0.01465098187327385, 0.008716126903891563, -0.0674104169011116, -0.006183144636452198, 0.04953689128160477, 0.018192976713180542, 0.03958144783973694, -0.07283449172973633, 0.07001076638698578, 0.031831879168748856, 0.3235403895378113, 0.05085519328713417, -0.2662697434425354, -0.08203066885471344, 0.0165033508092165, -0.0225446168333292, -0.07444852590560913, 0.024217279627919197, 0.1480993628501892, -0.12500834465026855, 0.07540807127952576, -0.0841100811958313, 0.0777493268251419, -0.04416468366980553, 0.004946528933942318, 0.12017876654863358, 0.11969776451587677, 0.005909114144742489, 0.07516135275363922, -0.17639242112636566, 0.18149740993976593, 0.018074525520205498, 0.10057837516069412, -0.051492732018232346, 0.04750869795680046, -0.0005896136863157153, 0.04373173415660858, 0.10841281712055206, 0.012017189525067806, -0.053542766720056534, -0.19188973307609558, -0.10950037837028503, 0.03781133145093918, 0.139766663312912, -0.0727108046412468, 0.07828284800052643, -0.06669971346855164, 0.001216001808643341, 0.01865432597696781, -0.05699155479669571, -0.1350599080324173, -0.15343786776065826, 0.02467147633433342, 0.006523122079670429, -0.03756207600235939, -0.09623881429433823, -0.0852329209446907, -0.04334240034222603, 0.12173648923635483, -0.0742320790886879, -0.042986731976270676, -0.1566188633441925, 0.029793834313750267, 0.143777534365654, -0.06484955549240112, 0.045049719512462616, 0.014957754872739315, 0.14323538541793823, 0.020194610580801964, -0.0401180274784565, 0.05763804167509079, -0.07387521117925644, -0.18827198445796967, -0.057529594749212265, 0.13721545040607452, 0.02183355763554573, 0.04626263305544853, -0.018393687903881073, 0.03803698718547821, -0.002372059738263488, -0.10752644389867783, 0.06032034009695053, 0.11071702092885971, 0.025790849700570107, 0.017381388694047928, -0.052060555666685104, 0.1148236095905304, -0.02599574625492096, -0.035334378480911255, 0.08819852769374847, 0.25935760140419006, -0.10474178194999695, 0.07166887074708939, 0.00983340386301279, -0.05558456480503082, -0.1442296802997589, 0.0017297746380791068, 0.12403963506221771, 0.03487970679998398, 0.06366100907325745, -0.17325061559677124, 0.07374612987041473, 0.12492005527019501, -0.027347758412361145, 0.061708007007837296, -0.290372759103775, -0.11061783879995346, 0.030355675145983696, 0.0851428434252739, -0.08433897793292999, -0.14451013505458832, -0.06274436414241791, -0.018610253930091858, -0.08572471141815186, 0.03838551789522171, -0.0723099410533905, 0.09798189997673035, -0.020791569724678993, 0.02302355132997036, 0.02512386254966259, -0.0358235165476799, 0.17778116464614868, -0.018518764525651932, 0.049929600208997726, -0.021519828587770462, 0.01655641943216324, 0.05964414030313492, -0.0779581367969513, 0.007721708156168461, -0.0825788751244545, 0.0794682577252388, -0.12462184578180313, -0.003901710966601968, -0.053228702396154404, 0.020810237154364586, -0.0662393569946289, -0.0475531667470932, -0.0424966886639595, 0.05964205041527748, 0.06839165836572647, -0.038327109068632126, 0.06362681835889816, 0.02596609853208065, 0.08907916396856308, 0.12326753884553909, 0.051042649894952774, 0.03203532472252846, -0.157696932554245, -0.018127206712961197, -0.01073373481631279, 0.06129911169409752, -0.12026309221982956, 0.016577132046222687, 0.12223176658153534, 0.05454052612185478, 0.10844683647155762, 0.025176996365189552, -0.05871386080980301, 0.008015472441911697, 0.04727160185575485, -0.0666792020201683, -0.1800532191991806, -0.02319747395813465, 0.014046290889382362, -0.14706005156040192, 0.006064438261091709, 0.10476724058389664, -0.03969049081206322, -0.01015650387853384, -0.0037907958030700684, 0.030906451866030693, -0.019695298746228218, 0.17423944175243378, 0.023275062441825867, 0.08246713131666183, -0.06656544655561447, 0.10144206881523132, 0.07422477006912231, -0.06455734372138977, 0.048016369342803955, 0.07082229107618332, -0.08263949304819107, -0.008668380789458752, 0.08006171882152557, 0.020145853981375694, 0.02158220112323761, -0.029632480815052986, -0.06353671103715897, -0.09655670821666718, 0.05100346729159355, -0.003594635520130396, 0.02074100822210312, -0.008341536857187748, -0.05188606679439545, 0.02572539448738098, -0.13071143627166748, 0.0873192623257637, 0.010543255135416985, 0.061687540262937546, -0.18366804718971252, 0.08232100307941437, 0.0031984501983970404, 0.018726155161857605, 0.0012909562792629004, 0.01722688414156437, -0.07806763797998428, -0.03261357545852661, -0.07630737125873566, -0.0005999861168675125, -0.02062823250889778, 0.013682258315384388, -0.01944497413933277, -0.011275961995124817, -0.034785546362400055, 0.03767049312591553, -0.050977569073438644, -0.09373050183057785, -0.01078538317233324, 0.07986084371805191, -0.11930789053440094, 0.01079732459038496, 0.03898872435092926, -0.10506384074687958, 0.0848451629281044, 0.007410438731312752, 0.04307807236909866, 0.011142093688249588, -0.058944493532180786, 0.023369114845991135, 0.01755213923752308, 0.02001640573143959, 0.042147912085056305, -0.11165044456720352, -0.02039492502808571, -0.05945315584540367, 0.004995713941752911, 0.011101646348834038, 0.02560962550342083, -0.12335827201604843, -0.031865693628787994, -0.05707279592752457, -0.05606904625892639, -0.057663898915052414, 0.04177315533161163, 0.07682691514492035, 0.01363006979227066, 0.12187211215496063, -0.06869523227214813, 0.04280601069331169, -0.2053670883178711, -0.03290747478604317, 0.007713342551141977, 0.017467210069298744, -0.0675274208188057, -0.016738463193178177, 0.08372344821691513, -0.04010125994682312, 0.10576187074184418, -0.033447764813899994, 0.07748287171125412, 0.042703039944171906, -0.06136833131313324, -0.02278079465031624, 0.012407040223479271, 0.20427264273166656, 0.061862871050834656, -0.009858480654656887, 0.0956491157412529, -0.036244846880435944, 0.031822193413972855, 0.09551955759525299, 0.12548241019248962, 0.16963349282741547, -0.011863687075674534, 0.05409466102719307, 0.05108917877078056, -0.1416195034980774, -0.1501016616821289, 0.1465211659669876, -0.0315585695207119, 0.10629613697528839, -0.0686730146408081, 0.13979332149028778, 0.09265682101249695, -0.21144917607307434, 0.022281218320131302, -0.057742588222026825, -0.1041441485285759, -0.10561687499284744, -0.04456564038991928, -0.07353613525629044, -0.12972542643547058, 0.014989055693149567, -0.09600900858640671, 0.03410962596535683, 0.09293504804372787, 0.0210861898958683, 0.025056732818484306, 0.13970741629600525, -0.021354490891098976, 0.012546221725642681, 0.07015008479356766, 0.046017736196517944, -0.000909544643945992, -0.05906134098768234, -0.05411018058657646, 0.032104793936014175, -0.05054044350981712, 0.09088850021362305, -0.05183554068207741, -0.021810660138726234, 0.056091174483299255, 0.021767456084489822, -0.06029559671878815, 0.032895397394895554, -0.0016047931276261806, -0.003030887572094798, 0.054224662482738495, 0.04643065109848976, -0.0024723822716623545, -0.0695456713438034, 0.2725455164909363, -0.07641040533781052, -0.016365107148885727, -0.13187547028064728, 0.16940221190452576, 0.015998534858226776, 0.0010140098165720701, 0.05234450101852417, -0.11191815137863159, -0.036647628992795944, 0.11681303381919861, 0.13924014568328857, -0.06140575930476189, -0.023158321157097816, 0.02045140415430069, -0.019495174288749695, -0.0643886923789978, 0.12683376669883728, 0.0787772461771965, 0.03386061638593674, -0.048445336520671844, 0.03908838331699371, -0.001963110873475671, -0.04026646167039871, -0.07436657696962357, 0.07367484271526337, 0.01886504888534546, 0.0161794014275074, -0.03964434564113617, 0.0712895393371582, 0.03543092682957649, -0.1447378396987915, 0.07277907431125641, -0.15929284691810608, -0.18839292228221893, -0.01753183826804161, 0.03980964422225952, -0.022502925246953964, 0.060254234820604324, -0.0017843334935605526, -0.012982836924493313, 0.16975843906402588, -0.017021620646119118, -0.04071721434593201, -0.11285580694675446, 0.07137885689735413, -0.07593928277492523, 0.23627938330173492, -0.012061760760843754, 0.051164135336875916, 0.11506396532058716, 0.010641196742653847, -0.1376773566007614, 0.015565470792353153, 0.08282727748155594, -0.10013698041439056, 0.022739078849554062, 0.1555064618587494, -0.04720205068588257, 0.12748198211193085, 0.05860716104507446, -0.09402312338352203, -0.010918213985860348, -0.02254103682935238, -0.007262950763106346, -0.08797216415405273, 0.025511635467410088, -0.05363193899393082, 0.16700366139411926, 0.218997523188591, -0.041029587388038635, -0.012612010352313519, -0.033150430768728256, 0.041794467717409134, 0.03437309339642525, 0.123933807015419, -0.0030796383507549763, -0.17697718739509583, 0.03857894986867905, -0.015100075863301754, 0.03110445849597454, -0.22140304744243622, -0.11476074904203415, 0.03637629374861717, -0.05475550517439842, -0.03744419664144516, 0.11287528276443481, 0.05315353348851204, 0.014867981895804405, -0.050493210554122925, -0.15417319536209106, -0.05335606634616852, 0.13279928267002106, -0.14487090706825256, -0.04436606541275978 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # t5-large-lora-4.72M-squad-model1 This model is a fine-tuned version of [t5-large](https://huggingface.co/t5-large) on the squad dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 47 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.1+cu121 - Datasets 2.15.0 - Tokenizers 0.15.0
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["varun-v-rao/squad"], "base_model": "t5-large", "model-index": [{"name": "t5-large-lora-4.72M-squad-model1", "results": []}]}
question-answering
varun-v-rao/t5-large-lora-4.72M-squad-model1
[ "transformers", "tensorboard", "safetensors", "t5", "question-answering", "generated_from_trainer", "dataset:varun-v-rao/squad", "base_model:t5-large", "license:apache-2.0", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T14:49:24+00:00
[]
[]
TAGS #transformers #tensorboard #safetensors #t5 #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-t5-large #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us
# t5-large-lora-4.72M-squad-model1 This model is a fine-tuned version of t5-large on the squad dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 47 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.1+cu121 - Datasets 2.15.0 - Tokenizers 0.15.0
[ "# t5-large-lora-4.72M-squad-model1\n\nThis model is a fine-tuned version of t5-large on the squad dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 47\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3", "### Training results", "### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0" ]
[ "TAGS\n#transformers #tensorboard #safetensors #t5 #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-t5-large #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us \n", "# t5-large-lora-4.72M-squad-model1\n\nThis model is a fine-tuned version of t5-large on the squad dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 47\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3", "### Training results", "### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0" ]
[ 81, 39, 6, 12, 8, 3, 90, 4, 33 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #t5 #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-t5-large #license-apache-2.0 #endpoints_compatible #text-generation-inference #region-us \n# t5-large-lora-4.72M-squad-model1\n\nThis model is a fine-tuned version of t5-large on the squad dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 47\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3### Training results### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0" ]
[ -0.08534107357263565, 0.16570717096328735, -0.0029720955062657595, 0.09835238754749298, 0.1071028932929039, 0.02016451396048069, 0.10924892127513885, 0.16228267550468445, -0.08945117890834808, 0.0850030854344368, 0.062015894800424576, 0.046411458402872086, 0.06021000072360039, 0.12777560949325562, -0.03418322280049324, -0.21603067219257355, 0.0036268436815589666, -0.018100079149007797, -0.06496426463127136, 0.10329379141330719, 0.11681367456912994, -0.09191033989191055, 0.08215320855379105, -0.01793692260980606, -0.11690380424261093, 0.03820158541202545, -0.017292428761720657, -0.06483928859233856, 0.09230945259332657, 0.01720321737229824, 0.06051667779684067, 0.013416655361652374, 0.11657741665840149, -0.2400560975074768, 0.0029479702934622765, 0.07741405814886093, 0.007585845421999693, 0.08405191451311111, 0.03829823061823845, 0.006893347017467022, 0.0922485738992691, -0.1773010641336441, 0.0994204506278038, 0.029034296050667763, -0.08154723048210144, -0.16632403433322906, -0.10992639511823654, 0.09097553789615631, 0.08672641962766647, 0.0877542495727539, 0.006056375801563263, 0.1601586937904358, -0.05093657970428467, 0.08561771363019943, 0.21908490359783173, -0.29496729373931885, -0.04581014811992645, 0.06300616264343262, 0.06230049580335617, 0.08399882167577744, -0.10339797288179398, 0.0018552164547145367, 0.05192051827907562, 0.01573266088962555, 0.09681572765111923, -0.00895662046968937, -0.08382391929626465, 0.013801679946482182, -0.12465200573205948, -0.04172792285680771, 0.20234601199626923, 0.06037764996290207, -0.035964857786893845, -0.11557335406541824, -0.059310927987098694, -0.08059965074062347, 0.001540919067338109, -0.07354793697595596, 0.04099234193563461, -0.05029468983411789, -0.04637724906206131, -0.07002697885036469, -0.08254565298557281, -0.06381756067276001, 0.004597408697009087, 0.049811337143182755, 0.05333033204078674, 0.01794295385479927, -0.037657514214515686, 0.08723311126232147, -0.018053287640213966, -0.1332162618637085, -0.03572479635477066, 0.010224051773548126, -0.08746536821126938, -0.05592309311032295, -0.014884015545248985, -0.05237007141113281, 0.011035382747650146, 0.1572117954492569, -0.06382317841053009, 0.036311231553554535, -0.009201751090586185, -0.005791467614471912, -0.024244965985417366, 0.12227941304445267, -0.0557430274784565, -0.040830206125974655, 0.01965332217514515, 0.08770096302032471, 0.03072587586939335, -0.00555967865511775, -0.07848384976387024, -0.03441164270043373, 0.09199550002813339, 0.0857769176363945, -0.008160494267940521, 0.023500211536884308, -0.025054922327399254, -0.02526027522981167, 0.01600203849375248, -0.1446382850408554, 0.035470739006996155, -0.030320601537823677, -0.0645114853978157, -0.07105199992656708, 0.0401366762816906, 0.0025675822980701923, -0.037011370062828064, 0.03788849711418152, -0.07017156481742859, -0.02403991110622883, -0.05580344796180725, -0.04984994977712631, 0.04518500715494156, -0.0710332840681076, -0.009672083891928196, -0.0682278573513031, -0.21261942386627197, -0.022291269153356552, 0.022927729412913322, -0.06381427496671677, -0.04646940529346466, -0.02386307157576084, -0.0681915432214737, 0.0027913975063711405, -0.010223638266324997, 0.09351852536201477, -0.033634547144174576, 0.07766273617744446, 0.011274627409875393, 0.040287408977746964, 0.04799850285053253, 0.03474294766783714, -0.09195646643638611, 0.03373171016573906, -0.10726779699325562, 0.05464952439069748, -0.08145291358232498, 0.020865362137556076, -0.13710583746433258, -0.09581194818019867, -0.003957892302423716, -0.03626840561628342, 0.05462818965315819, 0.12094515562057495, -0.17177848517894745, 0.002295898040756583, 0.17725308239459991, -0.09414657205343246, -0.14384028315544128, 0.11184442788362503, -0.04169408977031708, 0.039186690002679825, 0.06397084891796112, 0.1567654013633728, 0.10255024582147598, -0.159981369972229, -0.03843187168240547, 0.005984732415527105, 0.045077212154865265, 0.02242105081677437, 0.08072567731142044, -0.004064776934683323, 0.05659753829240799, 0.005582481622695923, -0.07682477682828903, -0.021921774372458458, -0.06552771478891373, -0.10060220211744308, -0.07671096175909042, -0.08480615168809891, 0.03140207380056381, 0.04343213886022568, 0.015633320435881615, -0.08126875013113022, -0.13024389743804932, 0.07651105523109436, 0.12449128925800323, -0.04986584186553955, 0.020123861730098724, -0.08373542129993439, 0.05627754330635071, -0.048379357904195786, -0.022145941853523254, -0.17385701835155487, -0.1291593462228775, 0.03640684857964516, -0.05374750867486, 0.02841564454138279, 0.018569419160485268, 0.06380034983158112, 0.057584743946790695, -0.07554667443037033, -0.025733554735779762, -0.09534809738397598, 0.007765965070575476, -0.08176477253437042, -0.1774960458278656, -0.03892671689391136, -0.03378364443778992, 0.11814652383327484, -0.21484977006912231, 0.026677049696445465, 0.027733799070119858, 0.15541304647922516, 0.046155378222465515, -0.045618198812007904, 0.011649125255644321, 0.0013320852303877473, -0.019929509609937668, -0.09569454938173294, 0.013554886914789677, -0.01279192790389061, -0.08295181393623352, -0.034830205142498016, -0.13763494789600372, 0.09762491285800934, 0.07908763736486435, 0.08560527861118317, -0.08596134930849075, -0.0011981732677668333, -0.06352502852678299, -0.03833330050110817, -0.08236366510391235, -0.03484414890408516, 0.12456414848566055, 0.00505084777250886, 0.10857025533914566, -0.089047871530056, -0.0900401771068573, 0.005895388312637806, 0.004424446728080511, -0.026377800852060318, 0.08108340948820114, 0.03353358432650566, -0.12239430099725723, 0.10725191980600357, 0.12031188607215881, -0.0017434136243537068, 0.12236274033784866, -0.07351729273796082, -0.10573834925889969, -0.045363232493400574, 0.03774723783135414, 0.0039318823255598545, 0.13088089227676392, -0.06157280504703522, 0.011683092452585697, 0.037627458572387695, 0.009294976480305195, 0.010741162113845348, -0.15554293990135193, -0.021916646510362625, 0.036186397075653076, -0.06463741511106491, 0.003908379003405571, -0.0036206389777362347, 0.018832292407751083, 0.097508005797863, 0.01583147794008255, 0.0000865399488247931, 0.026217468082904816, -0.013527686707675457, -0.09260688722133636, 0.16568423807621002, -0.09865441173315048, -0.16431432962417603, -0.11204183846712112, 0.07051724195480347, -0.03792700543999672, -0.03326890245079994, 0.028900472447276115, -0.0808233767747879, -0.05753428116440773, -0.11218705028295517, -0.015341001562774181, -0.019901877269148827, -0.011077312752604485, 0.046420611441135406, 0.03999628871679306, 0.08681695908308029, -0.1398712545633316, 0.020157916471362114, -0.0016293718945235014, -0.08859174698591232, -0.02331363968551159, 0.03880293667316437, 0.12671348452568054, 0.09400222450494766, -0.03216898813843727, 0.033146779984235764, -0.04666254669427872, 0.20239566266536713, -0.07775455713272095, 0.022276736795902252, 0.13037975132465363, 0.0075293793343007565, 0.059521809220314026, 0.12860606610774994, 0.01837364211678505, -0.08652449399232864, 0.03785808011889458, 0.06789218634366989, -0.022315489128232002, -0.2831334173679352, -0.027371680364012718, -0.02362552471458912, -0.03733326494693756, 0.07911238074302673, 0.07027965039014816, 0.049020130187273026, 0.037200361490249634, -0.030479129403829575, 0.0245626550167799, 0.004619965795427561, 0.08784156292676926, 0.09914584457874298, 0.025917410850524902, 0.07966756075620651, -0.05544288828969002, -0.031803227961063385, 0.0731450766324997, 0.039922893047332764, 0.2587493062019348, -0.023864956572651863, 0.14491496980190277, 0.029927019029855728, 0.15798655152320862, -0.05062238872051239, 0.03316335752606392, -0.0006532359402626753, 0.016871051862835884, 0.008131332695484161, -0.07860928028821945, 0.008367552421987057, 0.06415814161300659, -0.03924041613936424, 0.049329109489917755, -0.06567469984292984, 0.06284620612859726, 0.040789127349853516, 0.24360819160938263, 0.048874303698539734, -0.2694322466850281, -0.07259958237409592, 0.02600342221558094, -0.035107821226119995, -0.03828294575214386, 0.016776680946350098, 0.13630367815494537, -0.10533779114484787, 0.06728498637676239, -0.05767998844385147, 0.08175983279943466, -0.010628027841448784, -0.016020122915506363, 0.047108400613069534, 0.07488612085580826, -0.002785155549645424, 0.10577322542667389, -0.20413118600845337, 0.2147192358970642, 0.029336173087358475, 0.10157567262649536, -0.07684309035539627, 0.03798612579703331, -0.0021397012751549482, 0.06902731209993362, 0.16705071926116943, -0.004576015751808882, -0.07716930657625198, -0.13241411745548248, -0.08943363279104233, 0.02289370261132717, 0.11383377760648727, -0.049467459321022034, 0.08589141815900803, -0.052110739052295685, -0.009034717455506325, 0.05167023465037346, -0.045757029205560684, -0.15680594742298126, -0.12380636483430862, 0.029219772666692734, -0.0023302871268242598, -0.04158088192343712, -0.08920878916978836, -0.10142991691827774, -0.04390757530927658, 0.16447876393795013, -0.0149482237175107, -0.05258943885564804, -0.13942930102348328, 0.06349290162324905, 0.1327943652868271, -0.07114763557910919, 0.021417370066046715, 0.029793664813041687, 0.1301271766424179, 0.031369321048259735, -0.08691522479057312, 0.0587519109249115, -0.05887732654809952, -0.18298955261707306, -0.05311208218336105, 0.15777921676635742, 0.023578200489282608, 0.04780955612659454, 0.020185891538858414, 0.030263347551226616, 0.02002209611237049, -0.08388761430978775, 0.028340797871351242, 0.06370548158884048, 0.09640468657016754, 0.03655528277158737, -0.08077432960271835, 0.003989871591329575, -0.0430867075920105, -0.03252524882555008, 0.1305510401725769, 0.19924113154411316, -0.0944155678153038, 0.10185131430625916, 0.04281477630138397, -0.08057010918855667, -0.16995777189731598, 0.04569471254944801, 0.05842462182044983, 0.0019010360119864345, 0.09360085427761078, -0.13677406311035156, 0.10045125335454941, 0.08828097581863403, -0.02660708874464035, 0.032432153820991516, -0.3124876320362091, -0.13442251086235046, 0.05751396715641022, 0.11185616254806519, -0.020044123753905296, -0.15534964203834534, -0.048677507787942886, -0.00829352717846632, -0.1464463472366333, 0.113758385181427, -0.11618585139513016, 0.0760200098156929, -0.008522581309080124, 0.07565578818321228, 0.029214058071374893, -0.03919672593474388, 0.12252010405063629, 0.03692257031798363, 0.07353535294532776, -0.069601871073246, 0.004637778736650944, 0.13981324434280396, -0.07723795622587204, 0.10320434719324112, -0.05010807514190674, 0.09048004448413849, -0.15515361726284027, -0.023582404479384422, -0.05511242523789406, 0.05732893943786621, -0.06647193431854248, -0.06705890595912933, -0.06577780842781067, 0.06477886438369751, 0.08335856348276138, -0.04228030890226364, 0.09233614057302475, 0.033774472773075104, 0.09380579739809036, 0.10977347940206528, 0.10111575573682785, 0.03293278440833092, -0.10449560731649399, 0.0019113911548629403, -0.030942268669605255, 0.04126030206680298, -0.15422603487968445, 0.04638160020112991, 0.11936328560113907, 0.04687661677598953, 0.13375966250896454, 0.009609467349946499, -0.06772007793188095, -0.01526909600943327, 0.034005604684352875, -0.11871891468763351, -0.19487203657627106, -0.020731650292873383, -0.032106686383485794, -0.15667732059955597, 0.03352144733071327, 0.09399061650037766, -0.060964930802583694, -0.013331720605492592, -0.014289689250290394, 0.0447281152009964, -0.003263850463554263, 0.14949315786361694, 0.06345180422067642, 0.0633762776851654, -0.06295132637023926, 0.12192903459072113, 0.08517154306173325, -0.08657224476337433, 0.06349602341651917, 0.06987076997756958, -0.08113855868577957, -0.027309317141771317, 0.056450895965099335, 0.13688932359218597, -0.01690663769841194, -0.04646560549736023, -0.08961933851242065, -0.06957416236400604, 0.04497155174612999, 0.12185558676719666, 0.03498806431889534, -0.0016019893810153008, -0.0016635259380564094, 0.01810913160443306, -0.13262461125850677, 0.14205384254455566, 0.042426690459251404, 0.0673602819442749, -0.15357442200183868, 0.061780963093042374, -0.0004935213946737349, 0.04621497169137001, -0.02063111774623394, 0.044445011764764786, -0.07103241235017776, -0.023112228140234947, -0.10925369709730148, 0.006305788643658161, -0.03057326190173626, 0.009747395291924477, -0.024014592170715332, -0.08524369448423386, -0.03974039480090141, 0.059339530766010284, -0.052851468324661255, -0.059284407645463943, 0.018455615267157555, 0.06552886217832565, -0.1713264286518097, -0.03490860387682915, 0.027523113414645195, -0.0863666981458664, 0.09973372519016266, 0.025181297212839127, 0.020934976637363434, 0.021170267835259438, -0.08453749865293503, 0.01989627629518509, 0.013476371765136719, 0.041315920650959015, 0.0461353100836277, -0.127104252576828, -0.005454475060105324, -0.02152363769710064, 0.016985345631837845, 0.027112094685435295, 0.030401945114135742, -0.12324415147304535, -0.021173715591430664, -0.07480250298976898, -0.049429457634687424, -0.048726946115493774, 0.04575422406196594, 0.09664346277713776, 0.00048800004879012704, 0.15322303771972656, -0.07026452571153641, 0.05132612586021423, -0.22803276777267456, -0.025704093277454376, 0.009058937430381775, -0.022414235398173332, -0.08166548609733582, -0.014997496269643307, 0.06448917835950851, -0.07138145714998245, 0.11492493748664856, -0.009923405945301056, 0.08737843483686447, 0.05820247903466225, -0.03176353871822357, -0.0018549349624663591, -0.0009890764486044645, 0.18779721856117249, 0.028004799038171768, -0.01760641671717167, 0.0681341141462326, -0.05334610491991043, 0.04859921708703041, -0.009828361682593822, 0.15042570233345032, 0.17160312831401825, -0.014629197306931019, 0.04217277839779854, 0.09312758594751358, -0.09944779425859451, -0.14262565970420837, 0.09388407319784164, -0.01682193949818611, 0.08973406255245209, -0.04345795884728432, 0.12579306960105896, 0.14111612737178802, -0.1766306310892105, 0.041566140949726105, -0.05267536640167236, -0.09647310525178909, -0.11126583814620972, -0.07134857773780823, -0.09425140917301178, -0.10874393582344055, 0.028573008254170418, -0.13200871646404266, 0.03339478000998497, 0.0663505345582962, 0.006492839194834232, -0.0008931397460401058, 0.1658741533756256, -0.0298959631472826, 0.027167407795786858, 0.04450104385614395, 0.03143082931637764, 0.005789661779999733, -0.04062812030315399, -0.03165971860289574, 0.054442521184682846, 0.015235021710395813, 0.06219254434108734, -0.03218594565987587, 0.025229597464203835, 0.024390075355768204, -0.0219266414642334, -0.06838666647672653, 0.009302739053964615, 0.025489572435617447, 0.029397964477539062, 0.043492790311574936, 0.058592163026332855, 0.008137102238833904, -0.03627708554267883, 0.2602156400680542, -0.0670139417052269, -0.07142112404108047, -0.1319635659456253, 0.12509803473949432, 0.0386727936565876, -0.009732759557664394, 0.07880470156669617, -0.13680168986320496, 0.005531041882932186, 0.13764838874340057, 0.13691991567611694, -0.01279870793223381, -0.008537824265658855, -0.011837506666779518, -0.0070287687703967094, -0.04597974196076393, 0.07426105439662933, 0.1012137308716774, 0.041176874190568924, -0.04558075591921806, -0.013386132195591927, 0.003394196042791009, -0.02290203981101513, -0.06729000061750412, 0.08918119221925735, 0.002702519530430436, 0.010506861843168736, -0.009678822010755539, 0.08579658716917038, 0.03158843517303467, -0.20118126273155212, 0.04902885481715202, -0.18588343262672424, -0.17304909229278564, -0.007131559774279594, 0.10229462385177612, -0.028774557635188103, 0.023162024095654488, -0.0017840712098404765, -0.0028150840662419796, 0.1261029690504074, 0.005412922240793705, -0.09632871299982071, -0.07523242384195328, 0.08896005153656006, -0.08385968953371048, 0.24554549157619476, 0.0041204593144357204, 0.07457071542739868, 0.10537011176347733, -0.020889339968562126, -0.1580600142478943, 0.04207700490951538, 0.08973292261362076, -0.030492907389998436, 0.024123968556523323, 0.1562882959842682, -0.040560491383075714, 0.10711973905563354, 0.06189882382750511, -0.10466264933347702, -0.03405196964740753, -0.0029615091625601053, -0.013614063151180744, -0.10843590646982193, 0.02656327560544014, -0.06723392009735107, 0.1615280658006668, 0.17123021185398102, -0.05508654937148094, 0.024658111855387688, -0.0749015361070633, 0.03496575728058815, 0.06394772976636887, 0.06659771502017975, 0.016730470582842827, -0.1668039709329605, 0.027608806267380714, 0.01877935789525509, 0.03979424387216568, -0.23945017158985138, -0.1056874543428421, 0.06282366812229156, -0.04575882852077484, -0.08365850895643234, 0.10997273772954941, 0.11803865432739258, 0.029295191168785095, -0.02975533716380596, -0.1118549033999443, -0.04503076896071434, 0.1381874978542328, -0.16157642006874084, -0.03454818204045296 ]
null
null
diffusers
### Scenery Dreambooth model trained by ArminSalem following the "Build your own Gen AI model" session by NxtWave. Project Submission Code: 2254161038 Sample pictures of this concept: ![0](https://huggingface.co/ArminSalem/scenery/resolve/main/sample_images/xzg_(1).jpg)
{"license": "creativeml-openrail-m", "tags": ["NxtWave-GenAI-Webinar", "text-to-image", "stable-diffusion"]}
text-to-image
ArminSalem/scenery
[ "diffusers", "safetensors", "NxtWave-GenAI-Webinar", "text-to-image", "stable-diffusion", "license:creativeml-openrail-m", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
2024-02-10T14:53:28+00:00
[]
[]
TAGS #diffusers #safetensors #NxtWave-GenAI-Webinar #text-to-image #stable-diffusion #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us
### Scenery Dreambooth model trained by ArminSalem following the "Build your own Gen AI model" session by NxtWave. Project Submission Code: 2254161038 Sample pictures of this concept: !0.jpg)
[ "### Scenery Dreambooth model trained by ArminSalem following the \"Build your own Gen AI model\" session by NxtWave.\n\nProject Submission Code: 2254161038\n\nSample pictures of this concept:\n\n !0.jpg)" ]
[ "TAGS\n#diffusers #safetensors #NxtWave-GenAI-Webinar #text-to-image #stable-diffusion #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n", "### Scenery Dreambooth model trained by ArminSalem following the \"Build your own Gen AI model\" session by NxtWave.\n\nProject Submission Code: 2254161038\n\nSample pictures of this concept:\n\n !0.jpg)" ]
[ 73, 54 ]
[ "passage: TAGS\n#diffusers #safetensors #NxtWave-GenAI-Webinar #text-to-image #stable-diffusion #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n### Scenery Dreambooth model trained by ArminSalem following the \"Build your own Gen AI model\" session by NxtWave.\n\nProject Submission Code: 2254161038\n\nSample pictures of this concept:\n\n !0.jpg)" ]
[ -0.14063820242881775, 0.14938879013061523, -0.0013897860189899802, 0.03891326114535332, 0.05012296140193939, -0.047020748257637024, 0.17291851341724396, -0.0034986790269613266, -0.0035284992773085833, 0.02036232501268387, 0.15607069432735443, 0.07368086278438568, 0.019647452980279922, 0.14985094964504242, -0.05574525147676468, -0.15439757704734802, 0.030563831329345703, 0.0625372976064682, -0.00111645576544106, 0.06150165572762489, 0.059447258710861206, -0.0747774988412857, 0.12377536296844482, -0.035657912492752075, -0.14170636236667633, -0.013247494585812092, -0.03740526735782623, -0.04995523393154144, 0.10305073112249374, 0.05100759118795395, 0.029516849666833878, 0.10136941820383072, 0.031043609604239464, -0.05354999005794525, 0.04030903801321983, -0.007583029102534056, -0.06276257336139679, 0.0536002591252327, 0.038916558027267456, 0.07808327674865723, 0.16334977746009827, 0.10161519050598145, -0.07614083588123322, 0.03352109715342522, -0.08536436408758163, 0.004755339585244656, 0.031109711155295372, 0.11039023846387863, 0.12935909628868103, 0.0734105184674263, 0.021526208147406578, 0.08567019551992416, 0.0495024211704731, 0.1260381042957306, 0.09801525622606277, -0.22384142875671387, -0.1119418516755104, 0.1718728244304657, 0.08644727617502213, 0.02708721160888672, -0.06725116074085236, 0.1169818565249443, 0.09516045451164246, 0.0003443703753873706, 0.041442859917879105, -0.05515772104263306, 0.020368866622447968, -0.0797857791185379, -0.08434820175170898, 0.039184216409921646, 0.2313176542520523, 0.08600673079490662, -0.025856561958789825, -0.06907092034816742, -0.1009412631392479, 0.05688865855336189, -0.05640674754977226, -0.01948629319667816, -0.048354025930166245, 0.018315857276320457, -0.043773166835308075, -0.054224893450737, -0.14227823913097382, -0.027714988216757774, 0.0005964143201708794, 0.07610893249511719, -0.009525609202682972, 0.07423657923936844, -0.13624654710292816, 0.08389373123645782, 0.012816214002668858, -0.14058144390583038, 0.019088104367256165, -0.10151448100805283, 0.06763696670532227, 0.05523506924510002, 0.047419145703315735, -0.08348070830106735, 0.06825853884220123, -0.012397817336022854, 0.04037950932979584, -0.03286191076040268, 0.03154539689421654, 0.09848558902740479, 0.011306671425700188, -0.0641305148601532, -0.07198020070791245, -0.11380589008331299, 0.0374114103615284, -0.051640044897794724, 0.028163669630885124, -0.040359970182180405, -0.106087327003479, 0.008626671507954597, -0.06196032837033272, 0.03501928970217705, 0.03836310654878616, 0.029127947986125946, -0.0027819799724966288, -0.039271026849746704, 0.2235422134399414, 0.07445979118347168, -0.009576304815709591, -0.051433008164167404, 0.022022578865289688, 0.05399781093001366, 0.07130738347768784, 0.00159425288438797, 0.011133057996630669, 0.003003328572958708, -0.08311329782009125, -0.040145475417375565, -0.01594342663884163, -0.006086898501962423, 0.015463437885046005, -0.1216653436422348, 0.0740324929356575, -0.17285162210464478, -0.12145567685365677, 0.04839814454317093, 0.08633559197187424, -0.014486054889857769, -0.0707261711359024, -0.053070370107889175, -0.07487762719392776, 0.008189186453819275, -0.002992836758494377, -0.023684535175561905, -0.010810745880007744, 0.03116135112941265, -0.0031025062780827284, 0.10972535610198975, -0.2057703137397766, -0.01641431823372841, -0.06988000869750977, 0.033503077924251556, -0.0028063065838068724, 0.0005563330487348139, -0.01610797829926014, 0.10941678285598755, 0.0030403186101466417, -0.021287325769662857, 0.019587978720664978, -0.0014773450093343854, 0.001123499358072877, 0.1855553537607193, -0.13049207627773285, 0.013013754971325397, 0.16804811358451843, -0.13078616559505463, -0.19402389228343964, 0.0899532213807106, 0.03343898430466652, 0.09793722629547119, 0.04929433763027191, 0.08846817165613174, 0.045126497745513916, -0.21042270958423615, 0.021396616473793983, 0.026970401406288147, -0.1421448439359665, -0.1891455054283142, -0.007350361905992031, 0.1505318582057953, -0.07681023329496384, 0.02127215452492237, -0.07816866040229797, 0.11597428470849991, -0.09322480857372284, -0.03864312171936035, -0.03754008933901787, -0.10646311938762665, -0.056303393095731735, 0.015461064875125885, 0.026079948991537094, -0.0291531253606081, 0.03237835317850113, -0.14074471592903137, 0.02922775223851204, -0.024731891229748726, -0.010888021439313889, -0.08237450569868088, 0.045929428189992905, -0.08256964385509491, -0.012278273701667786, 0.012521577998995781, -0.06619217991828918, 0.03537524491548538, 0.13041752576828003, -0.024833189323544502, 0.16778087615966797, 0.05747988075017929, 0.10141757875680923, -0.03766649588942528, -0.0775858610868454, 0.07758311182260513, 0.001076522865332663, -0.016194701194763184, -0.17735803127288818, 0.07559973001480103, -0.06080079823732376, -0.06319665163755417, -0.19892434775829315, 0.011204260401427746, -0.023504024371504784, 0.10694828629493713, 0.059616412967443466, -0.0043918415904045105, 0.007641147822141647, -0.01984505169093609, -0.0988968163728714, -0.0049309260211884975, 0.06399621814489365, 0.04622189328074455, -0.07942767441272736, 0.19076915085315704, -0.09023676067590714, 0.19629496335983276, 0.06348786503076553, -0.06903751939535141, -0.026960495859384537, -0.011460382491350174, -0.07785625010728836, -0.01643628440797329, 0.00027030365890823305, 0.03341440483927727, -0.015157891437411308, -0.02326992340385914, 0.1347905844449997, -0.056137025356292725, 0.011357326060533524, 0.07523279637098312, -0.03999585285782814, -0.05395802855491638, 0.0793343260884285, 0.055016279220581055, -0.1715591996908188, 0.10345403850078583, 0.1157020702958107, 0.02486528269946575, 0.17525357007980347, 0.029018370434641838, -0.008808706887066364, -0.06684807687997818, 0.07107776403427124, 0.03147732466459274, 0.23764026165008545, -0.12371182441711426, 0.02608385495841503, 0.006959045305848122, -0.027117040008306503, 0.04539322853088379, -0.09905707836151123, -0.08179187774658203, 0.0077944668009877205, 0.0017594547243788838, 0.13660800457000732, 0.0990763008594513, -0.14432208240032196, 0.08773354440927505, -0.09539030492305756, -0.13811330497264862, 0.04117570444941521, -0.007479545194655657, -0.04640718549489975, 0.08509692549705505, -0.02437194436788559, -0.2788291573524475, -0.14423513412475586, -0.0878615602850914, -0.060838308185338974, -0.003685462288558483, 0.05892564728856087, -0.018998362123966217, -0.047987718135118484, -0.08286307007074356, -0.07735976576805115, -0.025183724239468575, 0.0034446914214640856, 0.02960912510752678, 0.034366823732852936, -0.03131077066063881, -0.044141482561826706, 0.02960861846804619, -0.027520593255758286, 0.046118300408124924, 0.123060442507267, 0.00109606864862144, 0.17766247689723969, 0.0753619447350502, 0.006586994044482708, -0.007136556785553694, 0.04020928964018822, 0.22032533586025238, -0.043335992842912674, 0.14094015955924988, 0.140453040599823, 0.06099887564778328, 0.06692778319120407, 0.15987446904182434, 0.04751204326748848, -0.09164932370185852, 0.041529711335897446, -0.07699970155954361, -0.12377159297466278, -0.10096953064203262, -0.07158752530813217, -0.05440926179289818, 0.11481890082359314, 0.016025738790631294, 0.08001668006181717, 0.13458965718746185, 0.1706732213497162, 0.021629229187965393, -0.017737528309226036, -0.06375846266746521, 0.10171680152416229, -0.03998909518122673, -0.04738525673747063, 0.015003016218543053, -0.06876347213983536, -0.060055311769247055, 0.10057076811790466, 0.053256671875715256, 0.12732785940170288, 0.035883087664842606, 0.022829843685030937, 0.08512593805789948, 0.150320902466774, 0.12852637469768524, 0.08811087906360626, -0.03277656435966492, -0.08668135851621628, 0.0002276510640513152, -0.07539829611778259, 0.10829295217990875, 0.09733863919973373, -0.07852131873369217, -0.019715679809451103, 0.056564051657915115, 0.07752328366041183, -0.038725752383470535, 0.10354812443256378, 0.12323086708784103, -0.2508062720298767, 0.007261200807988644, 0.012193383648991585, 0.060247596353292465, -0.1042860597372055, 0.038946449756622314, 0.26465851068496704, -0.008882184512913227, 0.07421738654375076, -0.050576865673065186, 0.08680728077888489, 0.05561321973800659, -0.003944547846913338, -0.05103013664484024, -0.0022872700355947018, 0.0043748365715146065, 0.025907672941684723, -0.14293743669986725, 0.19831623136997223, -0.022777365520596504, 0.04684392735362053, -0.00919387023895979, -0.04674613103270531, -0.0397929847240448, 0.18921761214733124, 0.2039906233549118, 0.029140247032046318, 0.0223556961864233, -0.011993310414254665, -0.11340425908565521, 0.023084981366991997, 0.013514896854758263, 0.033065080642700195, 0.008088999427855015, 0.08111423254013062, -0.05514474958181381, 0.00135017279535532, 0.04233679920434952, -0.18677447736263275, -0.11200389266014099, -0.020067937672138214, 0.24277514219284058, 0.06726318597793579, -0.02539687044918537, 0.014169509522616863, -0.04713661968708038, 0.15652206540107727, -0.1084517240524292, -0.04661247879266739, -0.078129343688488, -0.09938336908817291, -0.004479655530303717, -0.03372970223426819, 0.037276655435562134, -0.08710508793592453, 0.08415793627500534, -0.059932708740234375, -0.10078882426023483, 0.04251781851053238, -0.17442168295383453, -0.061100587248802185, -0.11803895980119705, 0.060076754540205, 0.05060355365276337, -0.049206383526325226, 0.030285578221082687, -0.0594177208840847, -0.06547363102436066, -0.10488709807395935, 0.024329597130417824, 0.023986293002963066, -0.05922219902276993, -0.06892647594213486, -0.047356028109788895, -0.03860273212194443, -0.011522849090397358, -0.08803374320268631, 0.08841150254011154, 0.25334954261779785, -0.0814175009727478, 0.05599992349743843, 0.21845921874046326, -0.03370104730129242, -0.21907539665699005, -0.12245234847068787, -0.0699823871254921, -0.028696002438664436, 0.0019431817345321178, -0.08976975828409195, 0.13490837812423706, -0.003972591366618872, -0.03391732648015022, 0.22086076438426971, -0.2750072479248047, -0.05738450214266777, 0.03089749999344349, 0.15325894951820374, 0.2973523736000061, -0.14978677034378052, -0.0323568694293499, -0.018504733219742775, -0.10660195350646973, 0.1529538780450821, 0.04635123535990715, 0.04435086250305176, -0.07535510510206223, 0.019817275926470757, -0.019801050424575806, -0.04849513620138168, 0.10377860814332962, -0.050273552536964417, 0.09902534633874893, -0.0797303169965744, 0.06699664890766144, 0.20464779436588287, -0.023802559822797775, 0.045078832656145096, -0.09081139415502548, 0.060953620821237564, -0.05474032461643219, -0.019797401502728462, -0.048521142452955246, 0.027837451547384262, -0.054906100034713745, -0.1116972491145134, -0.04564579948782921, -0.0013441889313980937, 0.008555032312870026, 0.03788677975535393, 0.007248756010085344, 0.004774775356054306, 0.009390733204782009, 0.17303122580051422, 0.012545266188681126, -0.08277872204780579, -0.024536773562431335, -0.08607804030179977, -0.05382510647177696, 0.11349093168973923, -0.017273494973778725, -0.03545167297124863, 0.10620173811912537, 0.01405181922018528, 0.032563112676143646, 0.019867150112986565, -0.012343418784439564, 0.054457105696201324, 0.11054637283086777, -0.16696226596832275, -0.16125145554542542, -0.07341565191745758, 0.21654854714870453, 0.12006817758083344, 0.12129959464073181, 0.11388456076383591, -0.10816314816474915, 0.0493810661137104, -0.05971338599920273, -0.012830273248255253, -0.005003783386200666, -0.008466409519314766, -0.017245350405573845, 0.02978929691016674, -0.07238193601369858, 0.024364925920963287, -0.0533960722386837, -0.10197966545820236, -0.05079283192753792, 0.08158661425113678, -0.08634010702371597, -0.08839502930641174, 0.04265039414167404, 0.1463303118944168, -0.19940948486328125, -0.08704623579978943, -0.006018958054482937, -0.06703589856624603, 0.04416020214557648, 0.08263026922941208, -0.003437579609453678, 0.01527540385723114, 0.02995501086115837, -0.03330851346254349, -0.07680970430374146, 0.024716675281524658, -0.022342499345541, 0.10767321288585663, -0.23736070096492767, -0.12903043627738953, -0.006837085820734501, 0.030221762135624886, -0.08649599552154541, -0.037704966962337494, -0.04423164576292038, 0.018112344667315483, -0.05592639371752739, 0.10018959641456604, -0.1296825408935547, -0.05139908939599991, -0.04643562436103821, -0.021168339997529984, -0.04134124144911766, 0.020111190155148506, -0.03412342816591263, 0.04897300526499748, 0.03998495638370514, -0.01951076276600361, -0.027627261355519295, -0.02652754634618759, -0.026547132059931755, -0.052111878991127014, 0.09057442098855972, -0.051540497690439224, -0.121290422976017, -0.04655526205897331, -0.26133012771606445, 0.005796317011117935, 0.1040843203663826, 0.013095760717988014, 0.006187965627759695, 0.09150176495313644, -0.013567733578383923, 0.044244859367609024, 0.023362571373581886, -0.047371670603752136, 0.001360428286716342, -0.11945132166147232, -0.04697068780660629, -0.0013562393141910434, -0.040971774607896805, -0.06906791031360626, -0.013726454228162766, 0.08975762873888016, -0.000574686739128083, 0.10188208520412445, -0.09883441776037216, 0.055089544504880905, -0.049536049365997314, 0.028775522485375404, 0.08749239891767502, -0.07235690951347351, -0.005077538080513477, -0.029922669753432274, -0.02789880707859993, 0.00017567820032127202, 0.09817018359899521, -0.06829769909381866, -0.20540037751197815, -0.02844339609146118, -0.12385956197977066, -0.05610102042555809, -0.009421661496162415, 0.3070743680000305, 0.017382998019456863, 0.0033298819325864315, -0.13453654944896698, 0.07347329705953598, 0.08386411517858505, 0.1022748053073883, -0.003173626959323883, 0.06614924222230911, 0.005157008301466703, 0.07702121138572693, 0.038078393787145615, 0.02152322605252266, -0.07635533064603806, -0.006236562971025705, -0.14989106357097626, 0.1540619283914566, -0.02521131932735443, 0.01741529256105423, 0.1870768666267395, 0.001865454250946641, -0.024735180661082268, 0.085823193192482, -0.02992790937423706, -0.027384815737605095, -0.23949331045150757, -0.07815693318843842, -0.10241241008043289, 0.032078105956315994, -0.05852658674120903, -0.0034423961769789457, -0.08659036457538605, 0.06846386939287186, -0.08683757483959198, 0.06283695995807648, 0.12724366784095764, -0.014997363090515137, 0.1260215789079666, 0.017655085772275925, -0.04065271094441414, 0.04973609745502472, 0.049362510442733765, -0.016163447871804237, -0.0025107392575591803, -0.04732444882392883, 0.062160927802324295, -0.004909321665763855, 0.058048803359270096, 0.04670637473464012, -0.05378849804401398, -0.02997903898358345, -0.025403302162885666, 0.010173920542001724, 0.07333064824342728, 0.01947731152176857, -0.008943886496126652, 0.019266393035650253, 0.11046309769153595, 0.0011535692028701305, -0.02069145441055298, -0.047495026141405106, 0.07654974609613419, -0.1314774453639984, 0.06978744268417358, -0.04436083883047104, 0.0086427703499794, -0.04124084860086441, 0.22910279035568237, 0.1431175172328949, -0.07865699380636215, -0.0014852599706500769, -0.07587140798568726, 0.006096833385527134, -0.08788558095693588, 0.10481026023626328, 0.016456009820103645, 0.2769347131252289, -0.06597760319709778, -0.05203752592206001, -0.11072257161140442, -0.034051090478897095, -0.0785800963640213, -0.08365166187286377, 0.029951676726341248, -0.03943697363138199, -0.11840207129716873, 0.07371119409799576, -0.1848495900630951, 0.023519674316048622, 0.09994617849588394, -0.01578694023191929, -0.007409917656332254, -0.04112511873245239, 0.11601768434047699, 0.03338706120848656, 0.027626939117908478, -0.09175612032413483, 0.04477248340845108, 0.056921131908893585, -0.029167737811803818, -0.09550070762634277, 0.07495300471782684, -0.016954338178038597, -0.1755625605583191, 0.15194305777549744, -0.022233769297599792, -0.016221610829234123, 0.06959214061498642, -0.04328372702002525, -0.14312000572681427, 0.0986003503203392, -0.030334580689668655, -0.06477095931768417, -0.012091653421521187, 0.1126071959733963, 0.02651968225836754, 0.01082325354218483, -0.004019658546894789, -0.0911661759018898, -0.04006195440888405, 0.10213720053434372, 0.015477826818823814, -0.07994373887777328, 0.08411804586648941, -0.0010286703472957015, 0.1018715500831604, -0.04804634302854538, -0.04558822885155678, -0.04287393018603325, -0.02581256814301014, 0.04861831292510033, 0.01807774417102337, -0.04200194031000137, 0.060149796307086945, -0.10304300487041473, -0.024158645421266556, 0.025570187717676163, 0.06216906011104584, -0.18879735469818115, -0.011490636505186558, -0.16494101285934448, -0.00032179433037526906, -0.04157843068242073, 0.010671103373169899, 0.20614409446716309, 0.01938595250248909, 0.01820138283073902, -0.024333102628588676, -0.00728516886010766, 0.027269242331385612, 0.013934334740042686, -0.13170816004276276 ]
null
null
diffusers
# DreamBooth trained by AutoTrain Text encoder was not trained.
{"tags": ["text-to-image", "diffusers", "autotrain"], "base_model": "stabilityai/stable-diffusion-xl-base-1.0", "instance_prompt": "alianawatson woman", "inference": true}
text-to-image
dominic1021/alianawatson
[ "diffusers", "text-to-image", "autotrain", "base_model:stabilityai/stable-diffusion-xl-base-1.0", "has_space", "region:us" ]
2024-02-10T14:56:47+00:00
[]
[]
TAGS #diffusers #text-to-image #autotrain #base_model-stabilityai/stable-diffusion-xl-base-1.0 #has_space #region-us
# DreamBooth trained by AutoTrain Text encoder was not trained.
[ "# DreamBooth trained by AutoTrain\n\nText encoder was not trained." ]
[ "TAGS\n#diffusers #text-to-image #autotrain #base_model-stabilityai/stable-diffusion-xl-base-1.0 #has_space #region-us \n", "# DreamBooth trained by AutoTrain\n\nText encoder was not trained." ]
[ 45, 19 ]
[ "passage: TAGS\n#diffusers #text-to-image #autotrain #base_model-stabilityai/stable-diffusion-xl-base-1.0 #has_space #region-us \n# DreamBooth trained by AutoTrain\n\nText encoder was not trained." ]
[ -0.02063869684934616, 0.12998254597187042, -0.00014558587281499058, 0.05282456427812576, 0.16523675620555878, 0.04722703993320465, 0.16625140607357025, 0.08092519640922546, -0.021600954234600067, 0.06268861889839172, 0.19911405444145203, -0.005327701102942228, 0.005592701490968466, 0.22998546063899994, -0.094501793384552, -0.15147385001182556, 0.05843960493803024, -0.017813973128795624, 0.08953600376844406, 0.04556926712393761, 0.01589704304933548, -0.08332102000713348, 0.06851272284984589, -0.1127990260720253, -0.21184474229812622, 0.06736689060926437, 0.028242893517017365, -0.08190154284238815, 0.023159906268119812, 0.057201284915208817, 0.11752432584762573, 0.05736266449093819, 0.06915528327226639, -0.09377864748239517, 0.030588991940021515, 0.09211067855358124, -0.037628743797540665, 0.060378964990377426, 0.002463718643411994, 0.007739691063761711, -0.03909904137253761, 0.01951049454510212, 0.05348891019821167, 0.033195290714502335, -0.09112479537725449, 0.09422965347766876, 0.008997537195682526, 0.05966416001319885, 0.005606517195701599, 0.1256808042526245, -0.02887202799320221, 0.0914452075958252, 0.0028242841362953186, 0.10286186635494232, 0.050214264541864395, -0.15577325224876404, -0.05811230465769768, 0.22586119174957275, 0.06323451548814774, 0.18434374034404755, -0.1056840792298317, 0.08215278387069702, 0.1282002329826355, 0.0043175057508051395, -0.024307064712047577, -0.0056144483387470245, -0.053464896976947784, -0.0875391811132431, -0.04101261869072914, -0.04863812029361725, 0.19171690940856934, 0.013884141109883785, -0.014532854780554771, -0.08809809386730194, -0.1092078685760498, -0.03936294838786125, 0.015471521764993668, 0.009576751850545406, -0.05643317475914955, 0.06334297358989716, -0.04036302492022514, -0.0881064385175705, -0.048688579350709915, -0.03869857266545296, -0.07886603474617004, 0.09238439798355103, -0.0456368625164032, 0.0745692178606987, -0.0938243567943573, 0.13909384608268738, -0.026598775759339333, -0.12820684909820557, 0.06501864641904831, -0.0971466526389122, 0.015486733056604862, 0.06505174934864044, -0.019916843622922897, -0.1562809944152832, 0.019901327788829803, 0.030637366697192192, 0.07526841759681702, 0.05189061909914017, -0.08258821815252304, 0.09015702456235886, 0.007376048713922501, 0.09042561054229736, -0.016077103093266487, -0.024903813377022743, 0.06223255768418312, 0.080438993871212, 0.023856146261096, -0.14336538314819336, -0.16565988957881927, 0.06790684908628464, -0.017159676179289818, 0.04283891245722771, 0.03642508387565613, -0.010275715962052345, -0.031149128451943398, -0.004403593484312296, 0.047221966087818146, -0.04838476702570915, 0.023466823622584343, -0.07434477657079697, -0.008917812258005142, 0.014335056766867638, 0.1431507170200348, 0.007567800115793943, -0.006044706329703331, -0.008012169972062111, -0.10112743824720383, -0.01249670796096325, -0.06397054344415665, -0.082596056163311, -0.05697616934776306, -0.11640746891498566, 0.03807840123772621, -0.16242456436157227, -0.1366284042596817, -0.010717466473579407, 0.012121928855776787, -0.08239061385393143, -0.0024879504926502705, -0.08431833982467651, -0.12462550401687622, 0.1450532078742981, -0.013907280750572681, -0.03597475588321686, 0.0006233238964341581, 0.06648663431406021, -0.010329908691346645, 0.10745283216238022, -0.17473040521144867, 0.01794232614338398, -0.07896706461906433, -0.0015359485987573862, -0.08321953564882278, 0.16549469530582428, -0.03203589841723442, 0.033024370670318604, -0.03292569890618324, 0.04207007214426994, 0.0021412093192338943, 0.008031118661165237, 0.05329214408993721, 0.15599198639392853, -0.19367799162864685, -0.04072578251361847, 0.0876203402876854, -0.08026987314224243, -0.011655561625957489, 0.041991058737039566, -0.022804416716098785, 0.047191135585308075, 0.005142057780176401, 0.15102070569992065, -0.07513030618429184, -0.1523657888174057, -0.00003674626350402832, 0.019653983414173126, -0.03947019204497337, 0.06174682825803757, -0.03899246081709862, 0.060578037053346634, -0.07573825865983963, 0.03253980353474617, -0.005597305484116077, 0.08249075710773468, -0.06469673663377762, -0.07055705785751343, -0.06726926565170288, -0.021799663081765175, 0.06577687710523605, 0.01678086258471012, 0.07544080168008804, -0.030378416180610657, -0.07784181833267212, 0.03869107738137245, 0.04462023451924324, -0.009920100681483746, -0.007784112356603146, -0.013205957598984241, -0.04446694254875183, -0.12920789420604706, 0.003658822737634182, -0.09591405093669891, -0.0857297033071518, 0.00785818975418806, 0.23912277817726135, 0.09514347463846207, 0.14679308235645294, 0.059998251497745514, 0.04194987192749977, -0.031193705275654793, -0.12705348432064056, -0.0008300838526338339, 0.029192514717578888, -0.08331938832998276, -0.09998124092817307, 0.0904180034995079, -0.09146905690431595, -0.004678551107645035, -0.1545001119375229, 0.007734695915132761, -0.07803455740213394, 0.15830396115779877, 0.028678199276328087, -0.031181402504444122, -0.03010755404829979, 0.0402386300265789, -0.09691616147756577, -0.1099129319190979, -0.0022663131821900606, 0.0153842493891716, -0.0945914015173912, 0.06970567256212234, -0.2405780851840973, 0.0574164092540741, 0.14391222596168518, -0.005025625228881836, -0.07321476936340332, 0.11765623092651367, 0.0489165261387825, -0.013706451281905174, -0.023128986358642578, -0.02168380096554756, 0.1244552806019783, -0.07626726478338242, 0.19949495792388916, -0.01798384077847004, 0.08187845349311829, 0.05062877759337425, -0.06974431127309799, -0.135806143283844, -0.000004087520210305229, -0.03837069496512413, -0.0334748737514019, 0.11700894683599472, 0.09331324696540833, -0.060808680951595306, 0.27977684140205383, 0.002255344530567527, -0.0019275352824479342, -0.03330899775028229, -0.014577753841876984, -0.0332055389881134, 0.12854062020778656, -0.012121065519750118, 0.00992091279476881, 0.015768490731716156, -0.014307437464594841, 0.01476898044347763, -0.09258662909269333, -0.015657516196370125, -0.029646404087543488, -0.0163404643535614, 0.1258670836687088, 0.016155531629920006, -0.035148244351148605, 0.07309972494840622, -0.04378744959831238, -0.0816405862569809, 0.11111503094434738, -0.022147411480545998, -0.0004421356425154954, 0.05905456468462944, -0.15857146680355072, -0.2807832360267639, -0.1459890753030777, 0.005951586179435253, -0.11860986053943634, 0.04109755903482437, 0.052975885570049286, -0.10799627006053925, -0.07004248350858688, -0.08202385157346725, -0.08629177510738373, -0.05557532608509064, 0.0011311533162370324, 0.11728531867265701, -0.06409677118062973, 0.05387398600578308, -0.06229059770703316, -0.00887343194335699, -0.013896237127482891, 0.0027349803131073713, 0.09634215384721756, 0.02155768871307373, 0.04409273341298103, 0.20931857824325562, -0.01992671564221382, 0.03497228026390076, -0.007471531629562378, 0.25480857491493225, -0.07225025445222855, 0.051100753247737885, 0.11487668752670288, 0.031045233830809593, 0.052618835121393204, 0.1828797161579132, -0.01034550741314888, -0.0642908588051796, 0.06494352221488953, -0.012484862469136715, -0.10492375493049622, -0.11105634272098541, -0.0924028679728508, -0.04872503876686096, -0.06293869018554688, 0.029581304639577866, 0.06633029878139496, 0.18465307354927063, 0.03403869643807411, -0.0085936663672328, 0.038062650710344315, -0.038405340164899826, 0.05253121256828308, 0.05000557377934456, -0.054350171238183975, 0.10506314784288406, -0.05272989347577095, -0.07878284156322479, 0.09704536944627762, 0.029444830492138863, 0.08175686746835709, -0.005787411238998175, -0.051862932741642, -0.054340463131666183, 0.05357728153467178, 0.12942302227020264, 0.016036581248044968, 0.0732298195362091, -0.037278078496456146, -0.04033561050891876, -0.043483830988407135, -0.012224663980305195, 0.08897408843040466, 0.023024603724479675, 0.013343557715415955, -0.06517297029495239, 0.09141328185796738, -0.0036450172774493694, 0.03365681692957878, 0.10284296423196793, -0.24468940496444702, 0.03720756992697716, 0.05340345576405525, 0.009430313482880592, -0.15917426347732544, -0.001802100450731814, 0.2596781551837921, -0.0778416246175766, -0.016604389995336533, -0.005158600863069296, 0.07767105102539062, 0.07948087900876999, -0.01405559852719307, -0.12727415561676025, 0.08470404893159866, -0.03762264549732208, -0.009994231164455414, -0.21587730944156647, 0.04233643785119057, 0.006741201039403677, 0.09690377861261368, -0.02572929486632347, 0.016345487907528877, 0.0344662107527256, 0.14141175150871277, 0.0716816708445549, 0.00973005685955286, -0.08598282933235168, -0.14106571674346924, -0.08402053266763687, -0.05161529779434204, 0.10742203146219254, 0.09498894214630127, -0.004010304808616638, -0.011004406958818436, 0.029761290177702904, 0.04038768634200096, -0.048020366579294205, -0.20780979096889496, -0.12313251197338104, 0.03342318534851074, 0.18468953669071198, 0.07250070571899414, -0.042261723428964615, -0.07773694396018982, 0.058913350105285645, 0.15853528678417206, -0.06002082675695419, -0.03646547347307205, -0.12438587844371796, -0.01314868126064539, 0.04682208597660065, -0.004984802100807428, 0.07632478326559067, -0.11283677071332932, 0.055372435599565506, -0.05680480971932411, -0.15995463728904724, 0.08369133621454239, -0.09573204070329666, -0.09156695753335953, -0.09880076348781586, -0.02600095607340336, -0.07628563791513443, -0.01809440366923809, 0.02631893940269947, 0.03644336014986038, -0.09317634254693985, -0.08042453974485397, 0.07387512177228928, 0.052659958600997925, -0.0790650025010109, 0.11336636543273926, 0.039935242384672165, -0.05932047963142395, 0.009086593985557556, -0.020160207524895668, 0.16297784447669983, 0.2692966163158417, -0.09637150168418884, 0.1332009732723236, 0.10272762179374695, -0.07975436747074127, -0.2972416281700134, -0.06331747770309448, -0.001001058961264789, 0.033033158630132675, -0.037056490778923035, -0.08421573042869568, 0.01754319854080677, -0.037301890552043915, -0.026686429977416992, 0.09380273520946503, -0.25594666600227356, -0.07236529886722565, 0.12090659141540527, 0.011188359931111336, 0.3046357333660126, -0.12652114033699036, -0.03758466988801956, -0.07161959260702133, 0.030579380691051483, 0.09310808032751083, 0.05593981221318245, 0.1552010029554367, -0.01064409501850605, 0.029015347361564636, 0.016381043940782547, -0.03504854813218117, 0.15569667518138885, -0.09976516664028168, 0.07290340214967728, -0.09811180084943771, 0.02065517008304596, 0.1682867556810379, -0.07824182510375977, 0.06025531142950058, -0.08820004016160965, 0.08328087627887726, -0.14803707599639893, 0.024164263159036636, -0.030000343918800354, 0.019950132817029953, 0.023836227133870125, -0.09545804560184479, -0.05183679237961769, -0.024305418133735657, 0.031683988869190216, 0.0011127261677756906, 0.008928169496357441, -0.03344632312655449, 0.021105246618390083, 0.31053033471107483, -0.045023828744888306, -0.08844760805368423, -0.032576143741607666, 0.0008607114432379603, -0.07616515457630157, 0.15518175065517426, -0.140009805560112, 0.016880689188838005, 0.08636961877346039, -0.028658051043748856, 0.19429416954517365, 0.04890631139278412, -0.034792251884937286, 0.06410761177539825, 0.08606549352407455, -0.17321881651878357, 0.023975208401679993, -0.08413522690534592, 0.03825248405337334, 0.07573363929986954, -0.08445089310407639, 0.1707473248243332, -0.07278440147638321, 0.0452447347342968, -0.039885539561510086, 0.022516414523124695, -0.02864324487745762, 0.07788124680519104, 0.05243882164359093, 0.03179828077554703, -0.08249194175004959, 0.1251235008239746, 0.038169246166944504, -0.00042698116158135235, 0.13369235396385193, 0.09562437236309052, -0.02339347079396248, -0.029987553134560585, -0.006221109069883823, 0.24116981029510498, -0.1580258458852768, -0.008135645650327206, -0.04209064692258835, -0.0893833190202713, -0.022283220663666725, 0.033660776913166046, 0.004361500032246113, 0.008071556687355042, -0.06307882070541382, -0.04562815651297569, -0.10188619047403336, 0.03915635868906975, 0.04616845026612282, 0.06768101453781128, -0.2191275805234909, 0.009082616306841373, 0.027556031942367554, 0.05952044948935509, -0.13306017220020294, -0.09101494401693344, -0.15259279310703278, 0.00039742272929288447, -0.13059686124324799, 0.06406794488430023, 0.061592768877744675, -0.04854949936270714, 0.035067036747932434, -0.043882932513952255, 0.0004143699479755014, 0.028861405327916145, -0.04535970091819763, -0.011117871850728989, 0.015505447052419186, 0.006177510134875774, -0.030567757785320282, -0.053487807512283325, -0.043063435703516006, -0.029680561274290085, 0.054787784814834595, 0.02104547619819641, -0.0758507251739502, -0.023473115637898445, -0.18298010528087616, -0.01812969706952572, 0.13245733082294464, 0.002356436103582382, -0.008200457319617271, 0.14597384631633759, -0.03255922719836235, 0.02350054867565632, 0.045941680669784546, 0.00834833923727274, 0.04824364185333252, -0.10032500326633453, -0.11206801235675812, -0.07519271969795227, -0.05291133001446724, -0.07905688136816025, 0.08345643430948257, 0.10799416899681091, 0.07442577183246613, 0.11548545211553574, -0.13865616917610168, 0.0669560506939888, -0.07766007632017136, -0.0069593568332493305, -0.02534155361354351, -0.07888507097959518, 0.010779611766338348, -0.010162390768527985, 0.045325834304094315, -0.0136204082518816, 0.14034360647201538, 0.09084946662187576, -0.13309991359710693, -0.0024138707667589188, -0.00929619837552309, -0.02631843276321888, -0.016799401491880417, 0.2551889717578888, 0.10145963728427887, -0.006956462282687426, -0.08942729234695435, 0.021860230714082718, 0.13579104840755463, 0.12235307693481445, 0.0031318129040300846, 0.015644969418644905, 0.024258719757199287, 0.16481736302375793, 0.003965459298342466, -0.016579868271946907, -0.059937287122011185, 0.03426060453057289, -0.10462383925914764, 0.12247732281684875, -0.11873367428779602, -0.14781181514263153, 0.10172251611948013, -0.02050800621509552, -0.03943636640906334, 0.0037147165276110172, -0.0770074650645256, -0.09716072678565979, -0.027703063562512398, -0.06800327450037003, -0.17266669869422913, 0.027315571904182434, -0.06229201331734657, 0.12446222454309464, 0.06566920876502991, 0.0076821851544082165, -0.07402129471302032, 0.09486519545316696, 0.02963947132229805, -0.0744946077466011, 0.11698131263256073, 0.006778121460229158, -0.004627263639122248, -0.0994558110833168, -0.04522183537483215, 0.07144024223089218, 0.1097252368927002, -0.0014914624625816941, 0.062032222747802734, 0.03687533363699913, 0.07175064831972122, -0.021852314472198486, -0.1358582228422165, 0.009430473670363426, 0.06657078862190247, -0.014820176176726818, 0.17750272154808044, 0.0526888370513916, 0.01400719489902258, -0.033916763961315155, 0.20015233755111694, -0.1060686707496643, -0.08181434869766235, -0.08442502468824387, 0.1438642144203186, -0.10308082401752472, 0.12064629793167114, -0.08840858936309814, -0.10166779905557632, -0.1078825369477272, 0.13028131425380707, 0.14787504076957703, -0.1705704927444458, -0.00973743386566639, -0.059736791998147964, -0.007106183096766472, -0.04775403439998627, 0.1790471076965332, 0.027036966755986214, 0.0724797174334526, -0.06618554890155792, 0.02645842544734478, -0.05293412134051323, -0.10053800046443939, -0.07210712134838104, -0.07806676626205444, 0.0033898563124239445, -0.046734727919101715, -0.1227606013417244, -0.054156865924596786, -0.1298540085554123, 0.07478922605514526, 0.13676925003528595, -0.09898433089256287, -0.036553580313920975, 0.0014106096932664514, 0.16058985888957977, -0.02301349863409996, -0.022042766213417053, -0.07060083746910095, 0.055989354848861694, 0.09736833721399307, -0.06496482342481613, -0.017015384510159492, -0.018570678308606148, -0.058492738753557205, -0.2682178318500519, 0.16817118227481842, -0.004298606421798468, 0.03949829190969467, 0.031070971861481667, 0.03424987196922302, -0.05919113755226135, 0.13132870197296143, -0.048278603702783585, -0.026828566566109657, -0.026347270235419273, 0.1921045184135437, -0.024049637839198112, 0.05515880510210991, 0.037085678428411484, -0.14625179767608643, -0.030122820287942886, 0.011054210364818573, -0.07361718267202377, 0.004916774109005928, -0.043997615575790405, -0.020118191838264465, 0.11343158781528473, 0.033629145473241806, -0.016268683597445488, 0.012032032944262028, -0.01385825127363205, 0.004258410073816776, -0.01866539753973484, -0.0066040013916790485, 0.027589673176407814, -0.1268121898174286, -0.026856685057282448, 0.0937948226928711, 0.038554634898900986, -0.2399676889181137, -0.056827764958143234, -0.20476673543453217, 0.0481138676404953, -0.07275717705488205, 0.13741889595985413, 0.15254093706607819, -0.023496365174651146, -0.007380692288279533, -0.12191148102283478, 0.015328208915889263, 0.04335465282201767, 0.00498174037784338, -0.033090222626924515 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
null
ahsenali/falcon-7b-med_dialogue-symptoms
[ "transformers", "safetensors", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
2024-02-10T15:05:01+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 31, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.06646376848220825, 0.2168014943599701, -0.00225935154594481, 0.023818302899599075, 0.1271018385887146, -0.001635765191167593, 0.04218708351254463, 0.13324736058712006, -0.020175931975245476, 0.11144465953111649, 0.046588581055402756, 0.09377603232860565, 0.09928803145885468, 0.18404334783554077, 0.04859916493296623, -0.2059975117444992, 0.007056170143187046, -0.09090408682823181, 0.014076028019189835, 0.1116579994559288, 0.13719257712364197, -0.10291384905576706, 0.08272874355316162, -0.04045208916068077, -0.02019004337489605, 0.00012576708104461432, -0.09259183704853058, -0.07032395154237747, 0.06885425746440887, 0.06264153122901917, 0.051234472543001175, 0.001456156256608665, 0.09140396863222122, -0.2864592671394348, 0.017265573143959045, 0.08406311273574829, 0.0027674848679453135, 0.06290827691555023, 0.07236549258232117, -0.07389893382787704, 0.11328595131635666, -0.08021481335163116, 0.13019037246704102, 0.08625296503305435, -0.062064990401268005, -0.23071379959583282, -0.07525765895843506, 0.0963398814201355, 0.12251301854848862, 0.06215599179267883, -0.022921854630112648, 0.15455181896686554, -0.06248689442873001, 0.012971068732440472, 0.1294165402650833, -0.11526761949062347, -0.05572471022605896, 0.061741601675748825, 0.11775490641593933, 0.10740239918231964, -0.14110268652439117, -0.0017287094378843904, 0.04900608956813812, 0.029121357947587967, 0.08589313924312592, 0.022661056369543076, 0.12003941088914871, 0.04652795568108559, -0.13695219159126282, -0.04037507623434067, 0.12011898308992386, 0.038862764835357666, -0.06446044892072678, -0.2168138176202774, -0.006778308190405369, -0.0601806715130806, -0.014732478186488152, -0.07019448280334473, 0.039128515869379044, -0.02470310963690281, 0.07317749410867691, -0.04465159401297569, -0.1063927412033081, -0.0421026237308979, 0.0892222449183464, 0.07748593389987946, 0.011527054943144321, -0.02519804798066616, 0.04627908393740654, 0.13455867767333984, 0.05402068421244621, -0.10399353504180908, -0.07017925381660461, -0.06942764669656754, -0.09420394152402878, -0.04035796597599983, 0.056760527193546295, 0.031942449510097504, 0.02665667235851288, 0.22703726589679718, 0.016653569415211678, 0.04155244305729866, 0.0224777739495039, 0.01032855175435543, 0.043662428855895996, 0.0955500528216362, -0.05303520709276199, -0.15660029649734497, -0.04072032496333122, 0.09077946096658707, -0.0027527001220732927, -0.036689214408397675, -0.03966725245118141, 0.03849169611930847, 0.06843466311693192, 0.13122352957725525, 0.07552056759595871, -0.017929591238498688, -0.04813180863857269, -0.030096933245658875, 0.23523783683776855, -0.1493375599384308, 0.04426715523004532, -0.02271856553852558, -0.01804111897945404, -0.03908449783921242, 0.03597262129187584, 0.022118929773569107, -0.000004518366949923802, 0.09706240892410278, -0.058981191366910934, -0.05378659814596176, -0.10168042778968811, -0.03272576630115509, 0.04088849574327469, -0.013975566253066063, -0.010589460842311382, -0.09025166928768158, -0.09490354359149933, -0.04766594246029854, 0.05537205561995506, -0.05123869329690933, -0.03770573064684868, 0.009465423412621021, -0.08151785284280777, -0.005444355774670839, -0.005417742300778627, 0.10699385404586792, -0.03222226724028587, 0.04445803165435791, -0.027600755915045738, 0.05225523188710213, 0.09919606149196625, 0.031576547771692276, -0.0773419588804245, 0.0561848059296608, -0.22559374570846558, 0.07503069192171097, -0.11481974273920059, 0.04335082694888115, -0.1704932004213333, -0.042439818382263184, 0.005444696638733149, 0.0139949731528759, 0.013206101022660732, 0.12720820307731628, -0.19255615770816803, -0.01654396951198578, 0.13260798156261444, -0.09212633967399597, -0.118110790848732, 0.07884611934423447, -0.029701577499508858, 0.1624738723039627, 0.04682036489248276, -0.027025915682315826, 0.09224298596382141, -0.16434773802757263, -0.07092688232660294, -0.00949116237461567, -0.01727987825870514, 0.12109188735485077, 0.07512219995260239, -0.05991523340344429, 0.046571120619773865, 0.02832140028476715, -0.038078423589468, -0.04424772411584854, -0.050857074558734894, -0.10884185880422592, -0.01070026308298111, -0.08987759798765182, 0.04065500199794769, -0.01250192429870367, -0.07916021347045898, -0.029885273426771164, -0.18612512946128845, -0.0030564051121473312, 0.10038342326879501, 0.0035033065360039473, -0.005652366206049919, -0.08666291832923889, 0.026358824223279953, -0.03112892620265484, -0.008404186926782131, -0.16764774918556213, -0.04399421438574791, 0.046902090311050415, -0.16094985604286194, 0.020117372274398804, -0.06413903087377548, 0.06334125250577927, 0.03641495108604431, -0.05590536445379257, -0.0248766727745533, -0.01730942726135254, 0.011945613659918308, -0.05083848536014557, -0.18994836509227753, -0.056277405470609665, -0.037882111966609955, 0.149809330701828, -0.25956398248672485, 0.032966937869787216, 0.051140617579221725, 0.14649195969104767, 0.00406361510977149, -0.05115427449345589, 0.01429014839231968, -0.05360214412212372, -0.054652128368616104, -0.06746816635131836, -0.006135428790003061, -0.027576493099331856, -0.05147203803062439, 0.019243421033024788, -0.1755700707435608, -0.021410830318927765, 0.09424154460430145, 0.12876708805561066, -0.1486445665359497, -0.018640631809830666, -0.048725154250860214, -0.06339836865663528, -0.0715010017156601, -0.07038594037294388, 0.10712739825248718, 0.0513901449739933, 0.04796046018600464, -0.07435787469148636, -0.07092321664094925, 0.02726263552904129, 0.006906150374561548, -0.03382374346256256, 0.08727246522903442, 0.05199531093239784, -0.09209315478801727, 0.0756213590502739, 0.1092359870672226, 0.07177663594484329, 0.09363535046577454, 0.01574566215276718, -0.11756632477045059, -0.028492970392107964, 0.036266472190618515, 0.02740776725113392, 0.1465986967086792, -0.05952361226081848, 0.04016614332795143, 0.04494241625070572, -0.04170418903231621, 0.022319864481687546, -0.08787637203931808, 0.024075502529740334, 0.025203049182891846, -0.0034381982404738665, 0.06284574419260025, -0.02525499276816845, -0.0050758360885083675, 0.07016654312610626, 0.047779910266399384, 0.04621000960469246, 0.009655474685132504, -0.01720241829752922, -0.1047825813293457, 0.16950392723083496, -0.0951867327094078, -0.269941508769989, -0.17632324993610382, 0.026197833940386772, 0.04035249724984169, -0.022378476336598396, 0.031619444489479065, -0.07056326419115067, -0.10630585998296738, -0.1060405746102333, -0.002429972169920802, 0.01714223250746727, -0.06364088505506516, -0.0741225928068161, 0.07348573952913284, 0.04382912442088127, -0.14902326464653015, 0.038552410900592804, 0.055694397538900375, -0.057955220341682434, -0.0233661737293005, 0.09118817001581192, 0.12397737801074982, 0.14583967626094818, -0.021366750821471214, -0.028626007959246635, 0.029004426673054695, 0.19620531797409058, -0.13469526171684265, 0.10371150821447372, 0.13814030587673187, -0.04545360431075096, 0.08360563963651657, 0.1560150384902954, 0.029186224564909935, -0.08317049592733383, 0.05044832453131676, 0.04082648828625679, -0.043159641325473785, -0.2666129767894745, -0.0534592866897583, 0.012832709588110447, -0.06255637854337692, 0.09786593168973923, 0.10183793306350708, 0.11542957276105881, 0.034910861402750015, -0.07166364789009094, -0.043925940990448, -0.0058974819257855415, 0.11737963557243347, -0.05490213260054588, -0.012639665976166725, 0.07686592638492584, -0.05086168646812439, 0.005355054512619972, 0.10266812145709991, 0.02973790094256401, 0.17442677915096283, 0.020399179309606552, 0.11231429129838943, 0.06195578724145889, 0.08633565157651901, 0.0007386076031252742, 0.02951662428677082, 0.05147615820169449, 0.017203815281391144, -0.002300140680745244, -0.10421168059110641, -0.006156572140753269, 0.1449710875749588, 0.028103826567530632, 0.029669636860489845, -0.0018948549404740334, -0.005003341939300299, 0.05121048167347908, 0.1746254414319992, -0.011592294089496136, -0.22072425484657288, -0.0845772922039032, 0.06936841458082199, -0.06218599155545235, -0.12968985736370087, -0.026130788028240204, 0.045467354357242584, -0.17519839107990265, 0.026703642681241035, -0.027433741837739944, 0.0919293761253357, -0.09345759451389313, -0.02221956104040146, 0.03687324374914169, 0.084866963326931, -0.014529162086546421, 0.08703910559415817, -0.14498743414878845, 0.11886418610811234, 0.02978132851421833, 0.09024628251791, -0.11081171780824661, 0.07909037172794342, -0.007550720125436783, 0.009180475026369095, 0.19379350543022156, -0.011335089802742004, -0.03514958545565605, -0.08774717897176743, -0.11210042238235474, -0.013537433929741383, 0.12687496840953827, -0.1243172138929367, 0.08773399889469147, -0.015198243781924248, -0.044079482555389404, 0.00937260314822197, -0.12100647389888763, -0.17273177206516266, -0.19628387689590454, 0.05585884302854538, -0.09575839340686798, 0.025643249973654747, -0.11914430558681488, -0.07089093327522278, -0.02952558360993862, 0.241120383143425, -0.1745356321334839, -0.06510113179683685, -0.1468164622783661, -0.046294767409563065, 0.1662203073501587, -0.04437198117375374, 0.0718095526099205, -0.0208172257989645, 0.20345525443553925, 0.005988610442727804, -0.004939318168908358, 0.06724198162555695, -0.08892562240362167, -0.16873881220817566, -0.06771010160446167, 0.1510489284992218, 0.11680185794830322, 0.04907919466495514, -0.002248800592496991, 0.0011772146681323647, -0.016943959519267082, -0.1137804463505745, -0.0033210667315870523, 0.16037839651107788, 0.03878779336810112, 0.025986969470977783, -0.05243593826889992, -0.08797456324100494, -0.06899320334196091, -0.06853509694337845, 0.06221301481127739, 0.19590823352336884, -0.10376439243555069, 0.1700313836336136, 0.147536963224411, -0.07305635511875153, -0.23175598680973053, 0.035342130810022354, 0.04983805492520332, 0.0014306638622656465, 0.04886869341135025, -0.18252557516098022, 0.10521943867206573, 0.019543392583727837, -0.05505957826972008, 0.13485197722911835, -0.1557481735944748, -0.1552847921848297, 0.0722852572798729, 0.03904085233807564, -0.22423844039440155, -0.1354004591703415, -0.09622503817081451, -0.05825018882751465, -0.14065024256706238, 0.06054598465561867, -0.002136280992999673, 0.015948504209518433, 0.03500790148973465, -0.0015643214574083686, 0.027123261243104935, -0.058935679495334625, 0.18609118461608887, -0.004065449349582195, 0.020676052197813988, -0.060264769941568375, -0.0478842556476593, 0.09839435666799545, -0.06130504235625267, 0.12208222597837448, 0.004057085141539574, 0.01594383642077446, -0.10362856835126877, -0.048314861953258514, -0.04328322783112526, 0.05154227837920189, -0.07548051327466965, -0.10070807486772537, -0.043625857681035995, 0.08841723203659058, 0.07005169242620468, -0.03383097052574158, 0.00549331633374095, -0.07189501076936722, 0.10019614547491074, 0.17795267701148987, 0.17573626339435577, 0.009926567785441875, -0.07241068035364151, 0.01677953451871872, -0.04142116755247116, 0.044231921434402466, -0.2513144314289093, 0.03756171092391014, 0.06098250672221184, 0.029438555240631104, 0.09217222779989243, -0.020435843616724014, -0.1820858269929886, -0.04050002992153168, 0.08094815909862518, -0.05452597141265869, -0.22617179155349731, -0.019085140898823738, 0.0954197570681572, -0.2020406424999237, -0.007372708059847355, 0.03995226323604584, -0.048725228756666183, -0.023169852793216705, 0.00010950004070764408, 0.06317184865474701, 0.002471912419423461, 0.09773622453212738, 0.0735151618719101, 0.09715340286493301, -0.08337292820215225, 0.10562895983457565, 0.10150538384914398, -0.09572599828243256, 0.03605884686112404, 0.06754924356937408, -0.05300498008728027, -0.043293699622154236, 0.03665391728281975, 0.033023297786712646, 0.005234600510448217, -0.060321882367134094, 0.013913018628954887, -0.036497246474027634, 0.044923391193151474, 0.08326134830713272, 0.03754979372024536, -0.013354414142668247, 0.06462216377258301, 0.03401726484298706, -0.10898099094629288, 0.10366570204496384, 0.01731540448963642, 0.04105307161808014, -0.08384523540735245, -0.019968897104263306, 0.035425446927547455, 0.030576206743717194, -0.01765924133360386, -0.02306121215224266, -0.02860277332365513, -0.01614218018949032, -0.14299540221691132, -0.023106401786208153, -0.07243485748767853, 0.006181265693157911, 0.014656842686235905, -0.031884219497442245, -0.011233693920075893, 0.02475680410861969, -0.06979699432849884, -0.07426341623067856, -0.006949664559215307, 0.09833318740129471, -0.15115703642368317, 0.008848577737808228, 0.06907843053340912, -0.11088496446609497, 0.08190931379795074, -0.008411259390413761, 0.016245156526565552, 0.022527478635311127, -0.15448406338691711, 0.05601610988378525, 0.0008648968650959432, 0.01916889287531376, 0.025886621326208115, -0.16471809148788452, 0.004104440100491047, -0.04661374166607857, -0.02149827405810356, -0.00004464812809601426, -0.02647159807384014, -0.12325995415449142, 0.06858719140291214, -0.015622655861079693, -0.035931166261434555, -0.02701525390148163, 0.0539589487016201, 0.07888586074113846, -0.027474910020828247, 0.10445091128349304, -0.008690856397151947, 0.04941811040043831, -0.16801609098911285, -0.02470702864229679, -0.04982255399227142, 0.019377702847123146, 0.009884213097393513, -0.007693959400057793, 0.04183054715394974, -0.00976533442735672, 0.21883612871170044, -0.05075952783226967, 0.1607085019350052, 0.05847611650824547, -0.017352959141135216, -0.0007513365126214921, 0.06180921941995621, 0.05997028574347496, 0.04658793285489082, 0.009480604901909828, 0.023740366101264954, -0.022450892254710197, -0.006695089396089315, -0.15932634472846985, 0.01890849508345127, 0.14999441802501678, 0.06301083415746689, 0.024745315313339233, 0.05866100639104843, -0.12775006890296936, -0.12135478109121323, 0.09311001747846603, -0.026755332946777344, 0.00928465835750103, -0.08245618641376495, 0.1358020007610321, 0.14980104565620422, -0.14000412821769714, 0.05256148427724838, -0.06134212389588356, -0.05217423290014267, -0.10388828068971634, -0.12032219022512436, -0.05887215584516525, -0.053666237741708755, 0.002330566756427288, -0.03760887682437897, 0.054546963423490524, 0.03344334661960602, -0.009351172484457493, -0.00022941511997487396, 0.13597318530082703, -0.019751882180571556, -0.0028988157864660025, 0.048313532024621964, 0.03693558648228645, 0.02373051457107067, -0.05275435373187065, 0.02940409444272518, 0.02539868652820587, 0.032232340425252914, 0.06546790152788162, 0.033412106335163116, -0.047448933124542236, 0.03804153576493263, -0.0025254099164158106, -0.11207924783229828, 0.019641218706965446, -0.00460948096588254, -0.0742158442735672, 0.1268945336341858, 0.0407399944961071, 0.010224059224128723, -0.03741471841931343, 0.24361543357372284, -0.06653323769569397, -0.06378097087144852, -0.13251738250255585, 0.10491154342889786, -0.0027236645109951496, 0.06476365029811859, 0.023412218317389488, -0.1284150779247284, 0.005243356805294752, 0.13858191668987274, 0.12181595712900162, 0.0045748427510261536, 0.009228081442415714, 0.0518609918653965, 0.0025186820421367884, -0.06998204439878464, 0.054019294679164886, 0.06992026418447495, 0.12919506430625916, -0.07847554981708527, 0.07680778950452805, 0.0006860480643808842, -0.08370215445756912, -0.02947772853076458, 0.11312682181596756, -0.0409729965031147, 0.03491825982928276, -0.047444481402635574, 0.10916327685117722, -0.05787910893559456, -0.29412412643432617, 0.02350960113108158, -0.09588567912578583, -0.15202060341835022, -0.018367812037467957, 0.05944539234042168, -0.02624768204987049, 0.018029648810625076, 0.06971040368080139, -0.06011629104614258, 0.20098382234573364, 0.0335683599114418, -0.07864278554916382, -0.0664360448718071, 0.04837050288915634, -0.06564252078533173, 0.2949807047843933, 0.008418165147304535, 0.02863333560526371, 0.10770907253026962, -0.03253700211644173, -0.18271861970424652, 0.010723991319537163, 0.1133992001414299, -0.08056149631738663, 0.08200647681951523, 0.19000613689422607, -0.012578671798110008, 0.1209007054567337, 0.05294662341475487, -0.047376248985528946, 0.04217283055186272, -0.03389401361346245, -0.051268599927425385, -0.10752558708190918, 0.058453381061553955, -0.05909625440835953, 0.15447644889354706, 0.10152646154165268, -0.05671518296003342, -0.004550917539745569, -0.05555408447980881, 0.04875178262591362, 0.01804669201374054, 0.12263146042823792, 0.02951994352042675, -0.1865430772304535, 0.032826557755470276, -0.01144319772720337, 0.10186848044395447, -0.25588861107826233, -0.08421015739440918, 0.08833149075508118, -0.011924264021217823, -0.05105875805020332, 0.10560628771781921, 0.057650718837976456, 0.04243382066488266, -0.043439045548439026, -0.10480839014053345, -0.02186836116015911, 0.14663739502429962, -0.1469624787569046, -0.025013303384184837 ]
null
null
diffusers
### omar_ai_interior_v1.1.0 Dreambooth model trained by davolu with [TheLastBen's fast-DreamBooth](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast-DreamBooth.ipynb) notebook Test the concept via A1111 Colab [fast-Colab-A1111](https://colab.research.google.com/github/TheLastBen/fast-stable-diffusion/blob/main/fast_stable_diffusion_AUTOMATIC1111.ipynb) Sample pictures of this concept:
{"license": "creativeml-openrail-m", "tags": ["text-to-image", "stable-diffusion"]}
text-to-image
davolu/omar-ai-interior-v1-1-0
[ "diffusers", "safetensors", "text-to-image", "stable-diffusion", "license:creativeml-openrail-m", "endpoints_compatible", "diffusers:StableDiffusionPipeline", "region:us" ]
2024-02-10T15:05:35+00:00
[]
[]
TAGS #diffusers #safetensors #text-to-image #stable-diffusion #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us
### omar_ai_interior_v1.1.0 Dreambooth model trained by davolu with TheLastBen's fast-DreamBooth notebook Test the concept via A1111 Colab fast-Colab-A1111 Sample pictures of this concept:
[ "### omar_ai_interior_v1.1.0 Dreambooth model trained by davolu with TheLastBen's fast-DreamBooth notebook\n\n\nTest the concept via A1111 Colab fast-Colab-A1111\n\nSample pictures of this concept:" ]
[ "TAGS\n#diffusers #safetensors #text-to-image #stable-diffusion #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n", "### omar_ai_interior_v1.1.0 Dreambooth model trained by davolu with TheLastBen's fast-DreamBooth notebook\n\n\nTest the concept via A1111 Colab fast-Colab-A1111\n\nSample pictures of this concept:" ]
[ 61, 57 ]
[ "passage: TAGS\n#diffusers #safetensors #text-to-image #stable-diffusion #license-creativeml-openrail-m #endpoints_compatible #diffusers-StableDiffusionPipeline #region-us \n### omar_ai_interior_v1.1.0 Dreambooth model trained by davolu with TheLastBen's fast-DreamBooth notebook\n\n\nTest the concept via A1111 Colab fast-Colab-A1111\n\nSample pictures of this concept:" ]
[ -0.1224161684513092, 0.004456889349967241, -0.0022144424729049206, 0.06676553934812546, 0.00915249902755022, -0.04060911387205124, 0.1413469910621643, -0.03216274082660675, 0.0035949349403381348, 0.05978820100426674, 0.14699915051460266, -0.00810061115771532, 0.0029571237973868847, 0.1663748323917389, -0.0544627420604229, -0.12237488478422165, 0.047360483556985855, 0.02360210195183754, -0.0543777234852314, 0.07799369096755981, 0.08167172223329544, -0.09177296608686447, 0.10849093645811081, -0.02639875002205372, -0.09322389960289001, -0.022926626726984978, -0.06372110545635223, -0.07775455713272095, 0.06197531521320343, 0.049098383635282516, 0.08399751037359238, 0.12644700706005096, 0.027766985818743706, -0.05382821708917618, 0.051845043897628784, -0.028214184567332268, -0.02623174898326397, 0.025342511013150215, -0.001804569736123085, 0.030393607914447784, -0.006346731912344694, 0.09225461632013321, 0.008653057739138603, 0.014939851127564907, -0.06452742218971252, 0.05687383562326431, 0.00817775633186102, 0.09908396750688553, 0.06137353181838989, 0.06596503406763077, 0.00946658942848444, 0.08432944864034653, 0.011424943804740906, 0.11326869577169418, 0.16086353361606598, -0.23940019309520721, -0.09655747562646866, 0.24723441898822784, 0.1263824850320816, -0.06456960737705231, -0.027265110984444618, 0.05478890612721443, 0.05572371184825897, 0.02333206869661808, -0.030346030369400978, -0.09036097675561905, -0.037810709327459335, -0.09769939631223679, -0.07477089017629623, 0.027957603335380554, 0.1849050521850586, 0.03184182196855545, -0.0748443454504013, -0.031033188104629517, -0.09153368324041367, 0.03643510863184929, -0.0672251433134079, -0.05549144744873047, -0.001915333210490644, -0.010650794953107834, -0.021247198805212975, -0.0013885913649573922, -0.11986803263425827, -0.06326093524694443, -0.07169925421476364, 0.15898698568344116, -0.03220420330762863, 0.04023600369691849, -0.050279002636671066, 0.1385304033756256, -0.01360991969704628, -0.1508166790008545, -0.020603418350219727, -0.11077523976564407, 0.07338167726993561, 0.0075212521478533745, -0.017022570595145226, -0.10673049837350845, 0.08682519942522049, 0.021847452968358994, 0.10294950753450394, -0.012179894372820854, 0.06718369573354721, 0.08093185722827911, -0.000007327462753892178, 0.018124155700206757, 0.0018863487057387829, -0.12554509937763214, 0.002096331911161542, 0.03482815623283386, 0.03235871344804764, -0.01912304200232029, -0.09198810160160065, -0.0006494690896943212, -0.016950180754065514, -0.0033120608422905207, 0.008960085920989513, 0.017671337351202965, -0.058621179312467575, -0.013967623934149742, 0.09990936517715454, -0.021081916987895966, -0.0289713554084301, -0.0524258092045784, -0.042677976191043854, -0.023203525692224503, 0.144582599401474, -0.022199496626853943, 0.013521057553589344, 0.09923109412193298, -0.06346514075994492, -0.010558178648352623, -0.020085280761122704, -0.032938919961452484, -0.006086994428187609, -0.05888356640934944, 0.055792465806007385, -0.16629736125469208, -0.17592036724090576, -0.008300050161778927, 0.0852467268705368, -0.050706181675195694, -0.005163198336958885, -0.026983914896845818, -0.11143707484006882, -0.009116404689848423, 0.027713632211089134, -0.02122281864285469, -0.026990262791514397, 0.04723444953560829, 0.0728951171040535, 0.10301525890827179, -0.10489178448915482, -0.03083026222884655, -0.09740222245454788, 0.05085514485836029, -0.11088307201862335, 0.0025113634765148163, -0.07526233792304993, 0.11660094559192657, -0.004785319324582815, -0.037246279418468475, 0.009785325266420841, 0.039312977343797684, 0.012577490881085396, 0.2049342691898346, -0.1554419994354248, 0.01300401333719492, 0.13593336939811707, -0.17342060804367065, -0.24341751635074615, 0.0690811425447464, 0.027586596086621284, 0.1238718032836914, 0.01771238073706627, 0.008357708342373371, 0.07536567747592926, -0.319415807723999, -0.02989319898188114, 0.023656057193875313, -0.07984211295843124, -0.09346266090869904, 0.03794023022055626, 0.13031290471553802, 0.008370059542357922, 0.02397823892533779, -0.02183212712407112, 0.05558951199054718, -0.07575096935033798, -0.0521140992641449, -0.06072581186890602, -0.05283376947045326, -0.037920281291007996, 0.005738032516092062, 0.018767787143588066, -0.05593064799904823, 0.008290480822324753, -0.053681645542383194, 0.00529684079810977, 0.027986792847514153, -0.0474817268550396, -0.10622355341911316, 0.09218326210975647, -0.1012527197599411, -0.020326867699623108, -0.02691744826734066, -0.0784263163805008, -0.006533000152558088, 0.1469673365354538, -0.03477063775062561, 0.15946465730667114, 0.08951088786125183, 0.0965704619884491, -0.004682749975472689, -0.0603497140109539, 0.04821426793932915, 0.03478989377617836, -0.033184610307216644, -0.14063803851604462, 0.09970615059137344, -0.07951374351978302, -0.005051645915955305, -0.11106576770544052, 0.026063738390803337, 0.06238259747624397, 0.1627163589000702, 0.08080455660820007, 0.018504353240132332, 0.02991374023258686, -0.00769343227148056, -0.03530016914010048, -0.06362445652484894, 0.03264031186699867, 0.03506677597761154, -0.03227189555764198, 0.11123477667570114, -0.0884079560637474, 0.30967795848846436, 0.10633783042430878, 0.008571437560021877, -0.07484643906354904, -0.03293410316109657, -0.05122413858771324, -0.016869818791747093, -0.02243223413825035, 0.0556006096303463, 0.03248272463679314, -0.01791207306087017, 0.13842639327049255, -0.0446782223880291, 0.027363630011677742, 0.058440662920475006, -0.07287564128637314, -0.0671655684709549, 0.06395328044891357, -0.046165820211172104, -0.07566550374031067, 0.0697302296757698, 0.17784804105758667, -0.029895106330513954, 0.14828427135944366, -0.0019794439431279898, 0.009829150512814522, -0.09020446985960007, 0.05607941746711731, -0.001630749087780714, 0.23912584781646729, -0.08914639800786972, 0.01738942228257656, 0.01312727015465498, -0.013066725805401802, 0.023801622912287712, -0.07123618572950363, -0.03171483799815178, 0.04930679500102997, 0.00624141376465559, 0.2009178251028061, 0.07596724480390549, -0.1151956245303154, 0.03773833438754082, -0.07180366665124893, -0.18596875667572021, 0.03130745142698288, -0.0007972284802235663, 0.009642168879508972, 0.12271462380886078, -0.024879205971956253, -0.2436034381389618, -0.0978265032172203, -0.09479445964097977, -0.0073300618678331375, -0.01111509557813406, 0.051219381392002106, 0.027274513617157936, -0.07103726267814636, -0.09713982045650482, 0.003637698944658041, 0.02933374047279358, 0.02545100823044777, 0.06648679822683334, 0.027687126770615578, -0.0439487062394619, -0.028337236493825912, -0.016386598348617554, -0.0230234507471323, 0.10242975503206253, 0.14888201653957367, -0.016386592760682106, 0.11604258418083191, 0.07339268922805786, -0.010650611482560635, 0.0018788964953273535, 0.031043458729982376, 0.28544434905052185, -0.04451579228043556, 0.10524984449148178, 0.176405131816864, 0.07729126513004303, 0.02651270106434822, 0.19659875333309174, 0.019942518323659897, -0.07344361394643784, 0.07536474615335464, -0.08892829716205597, -0.07773588597774506, -0.03568926081061363, -0.09963864088058472, -0.021760428324341774, 0.08194328099489212, -0.022672414779663086, 0.062239669263362885, 0.019586700946092606, 0.16100196540355682, 0.08097030967473984, 0.02134045958518982, -0.044998444616794586, 0.0814528688788414, 0.10491456836462021, -0.07685859501361847, 0.03265165165066719, -0.052662596106529236, -0.10188331454992294, 0.0926918089389801, 0.02985159307718277, -0.0015378051903098822, -0.04296734184026718, -0.09770266711711884, 0.053119197487831116, 0.07533710449934006, 0.1172291487455368, 0.10401201248168945, 0.006188335828483105, -0.10576483607292175, -0.030165959149599075, -0.09623903036117554, 0.030975280329585075, 0.07545451819896698, -0.07978583127260208, 0.016778651624917984, 0.030282720923423767, 0.13117453455924988, -0.01263240072876215, 0.02770981192588806, 0.16131281852722168, -0.26603594422340393, -0.04713597521185875, -0.024069026112556458, 0.07681219279766083, -0.10396643728017807, 0.016415221616625786, 0.2239614576101303, 0.0053779855370521545, -0.0011014655465260148, -0.08693983405828476, 0.050188664346933365, 0.0909881517291069, 0.00616246834397316, -0.07906115055084229, -0.0038158048409968615, -0.023362040519714355, 0.012254341505467892, -0.21644346415996552, 0.0942293182015419, -0.020427653566002846, 0.08674859255552292, 0.022301239892840385, -0.004893817938864231, 0.01168028637766838, 0.18324941396713257, 0.17451262474060059, -0.02600913681089878, 0.08254469931125641, -0.0008436607313342392, -0.15528197586536407, 0.008790690451860428, 0.03538326174020767, 0.07510962337255478, 0.04182802140712738, 0.04949470981955528, -0.02493560127913952, 0.0046687196008861065, 0.05226469784975052, -0.17083615064620972, -0.0353027880191803, 0.042975299060344696, 0.10600168257951736, 0.06157299503684044, -0.0827527716755867, -0.05991277098655701, 0.017883537337183952, 0.15709806978702545, -0.1344846487045288, -0.04128545522689819, -0.10051704198122025, -0.07714777439832687, 0.04732506722211838, -0.0161475520581007, 0.05711745098233223, -0.09934409707784653, 0.04607725888490677, -0.04702290520071983, -0.06430341303348541, 0.059118516743183136, -0.17304448783397675, -0.09880253672599792, -0.14652477204799652, 0.027881328016519547, -0.054062265902757645, -0.016006020829081535, 0.025407465174794197, -0.024599190801382065, -0.09071800112724304, -0.0785769671201706, -0.013324839994311333, 0.007752462290227413, -0.05122903734445572, -0.004545352887362242, 0.01384615246206522, -0.04202213138341904, 0.034157563000917435, -0.009280884638428688, 0.04370567947626114, 0.2686016857624054, -0.030424917116761208, 0.051643773913383484, 0.1507052332162857, -0.00655194791033864, -0.24400100111961365, -0.140359565615654, -0.05185229703783989, 0.03261551633477211, -0.04851720482110977, -0.007669789716601372, 0.1970830261707306, -0.008740853518247604, -0.04837570711970329, 0.2275247424840927, -0.2750934958457947, -0.08683362603187561, 0.14620724320411682, 0.13217630982398987, 0.34358933568000793, -0.12903808057308197, -0.04072735458612442, -0.027037741616368294, -0.26549118757247925, 0.07824279367923737, 0.05023859441280365, 0.06953916698694229, -0.094504714012146, -0.009128021076321602, -0.011927114799618721, -0.07211680710315704, 0.12807466089725494, -0.10046190768480301, 0.07334642857313156, -0.13378970324993134, 0.027860887348651886, 0.13343940675258636, -0.03911328315734863, 0.07853621989488602, -0.05791270360350609, 0.09920929372310638, 0.00025003025075420737, -0.03982699289917946, 0.0010968713322654366, 0.06933775544166565, -0.06615898758172989, -0.10696954280138016, -0.05843740329146385, 0.05280618742108345, -0.03359316289424896, -0.026433760300278664, -0.08209332823753357, -0.009778604842722416, -0.08123081922531128, 0.22258451581001282, -0.031168567016720772, -0.10794684290885925, -0.04601659998297691, -0.009156688116490841, -0.04140310734510422, 0.09876960515975952, -0.0006331963813863695, -0.04390260577201843, 0.1699601262807846, 0.00906446110457182, 0.07778257876634598, 0.04352240264415741, -0.023091834038496017, 0.0011723608477041125, 0.10307936370372772, -0.16380523145198822, -0.028543855994939804, -0.04726402461528778, 0.12826550006866455, 0.03479067608714104, 0.009100658819079399, 0.16023491322994232, -0.132095068693161, 0.06697408854961395, -0.044255226850509644, -0.016171423718333244, -0.006428547203540802, 0.11581943184137344, 0.005592662375420332, 0.04933767020702362, -0.03942551463842392, 0.05223029479384422, -0.06906124949455261, -0.14416278898715973, -0.07640938460826874, 0.045391183346509933, -0.097920261323452, -0.0746569037437439, 0.07397361844778061, 0.19066207110881805, -0.13175804913043976, -0.029367197304964066, -0.13598360121250153, -0.13927921652793884, 0.04398298263549805, 0.180974543094635, 0.07698395103216171, 0.05935748293995857, 0.02615976333618164, -0.0526663176715374, -0.007605458609759808, 0.0975855216383934, 0.04802166670560837, 0.08880485594272614, -0.22958824038505554, -0.05920752137899399, -0.049762267619371414, 0.037965089082717896, -0.09464097768068314, -0.016135217621922493, -0.10075069963932037, 0.0007549808942712843, -0.09575209766626358, 0.12725615501403809, -0.06233645975589752, -0.04194634035229683, 0.0028073180001229048, -0.007964201271533966, -0.020967375487089157, 0.010920667089521885, -0.03155539557337761, 0.0514955148100853, -0.0015568475937470794, -0.0188068225979805, -0.04699647054076195, -0.06086701154708862, 0.018782980740070343, -0.06214933097362518, 0.056849826127290726, -0.020678505301475525, -0.11166631430387497, -0.06847740709781647, -0.23753343522548676, 0.007974930107593536, 0.12749651074409485, -0.016534827649593353, 0.029058106243610382, 0.04000324010848999, -0.02777211181819439, 0.023497505113482475, 0.02222602069377899, -0.0017873927718028426, 0.11120029538869858, -0.10746900737285614, -0.07089214771986008, -0.05246223881840706, -0.025843506678938866, -0.08615715056657791, -0.002711558947339654, 0.11308076977729797, 0.07027779519557953, 0.14266404509544373, -0.1302008032798767, 0.041733235120773315, -0.04991263523697853, 0.0005098825786262751, 0.06968630105257034, -0.0543014258146286, 0.02828449197113514, 0.0034347758628427982, -0.004349921364337206, 0.005718943662941456, 0.11508917063474655, 0.014260723255574703, -0.1773248165845871, 0.0005504448781721294, -0.13307172060012817, -0.002237000036984682, 0.018501535058021545, 0.19571152329444885, 0.03002564050257206, 0.03315720707178116, -0.13978156447410583, 0.06312213093042374, 0.12389245629310608, 0.07947355508804321, 0.0319889560341835, 0.12356030941009521, -0.010516524314880371, 0.1739589273929596, 0.007328872103244066, 0.0348520427942276, 0.010186723433434963, -0.005161161534488201, -0.1224551871418953, 0.12883880734443665, -0.041712697595357895, -0.03802744299173355, 0.08901374042034149, 0.02234828844666481, -0.054144032299518585, -0.008267496712505817, -0.05048224702477455, -0.013376149348914623, -0.04456724226474762, -0.09081752598285675, -0.08407402038574219, 0.03508473187685013, -0.06031119450926781, -0.04064411297440529, -0.0024529313668608665, 0.056237004697322845, -0.011789420619606972, 0.15918487310409546, -0.05063394457101822, 0.0066738915629684925, 0.12589430809020996, -0.011575007811188698, -0.05417848750948906, 0.0239406805485487, 0.0452599972486496, -0.05965939164161682, 0.06817017495632172, -0.059036944061517715, 0.046673472970724106, 0.00041949236765503883, 0.026530073955655098, 0.0470472052693367, -0.05832456424832344, -0.02801092341542244, 0.010571936145424843, 0.03162246569991112, 0.05920126289129257, 0.014609710313379765, 0.025155717507004738, 0.01293738279491663, 0.1655249148607254, -0.048216577619314194, -0.18866831064224243, -0.08927930146455765, 0.009194055572152138, -0.09748191386461258, 0.08236421644687653, -0.03914891183376312, -0.03208243474364281, -0.03309456631541252, 0.15593720972537994, 0.13337451219558716, -0.1297912299633026, 0.0013509822310879827, -0.006726031191647053, 0.0025020495522767305, -0.06550134718418121, 0.030071375891566277, 0.026030823588371277, 0.28017377853393555, -0.08586570620536804, -0.06032885983586311, -0.09302771836519241, -0.08283363282680511, -0.015593769028782845, -0.12972921133041382, 0.03788027912378311, -0.010309715755283833, -0.1159377470612526, 0.06550634652376175, -0.1449187695980072, -0.04896778613328934, 0.22332227230072021, -0.0738975927233696, -0.051499541848897934, -0.055714040994644165, 0.13093477487564087, 0.040817685425281525, 0.04849892109632492, -0.0958276093006134, -0.004462071694433689, 0.017817972227931023, -0.04767473787069321, -0.09439833462238312, 0.07859963923692703, 0.0005187169881537557, -0.14538529515266418, 0.1827031373977661, 0.009523973800241947, 0.062002237886190414, 0.07115554809570312, -0.019805312156677246, -0.1073644608259201, 0.09898906201124191, -0.011987509205937386, -0.09983043372631073, 0.001897504203952849, 0.07561223953962326, 0.030513105913996696, 0.02098657563328743, 0.005987571086734533, -0.1349770575761795, -0.03256518021225929, 0.1348823457956314, -0.012559459544718266, -0.1391962468624115, 0.0819491446018219, -0.009776494465768337, 0.08086024224758148, 0.044207509607076645, -0.0446942113339901, 0.008159779012203217, -0.02229287475347519, 0.09445963054895401, 0.004661284852772951, -0.058653365820646286, 0.06677227467298508, -0.05996456742286682, 0.019078223034739494, -0.02537442371249199, -0.03434893488883972, -0.23166126012802124, -0.08994938433170319, -0.14799726009368896, 0.015078629367053509, -0.0065470049157738686, 0.0600280836224556, 0.2033112347126007, 0.05938434973359108, -0.0003695210616569966, 0.04643760621547699, -0.015968386083841324, 0.02388952672481537, -0.022660408169031143, -0.14398042857646942 ]
null
null
null
A WASM version of Tower
{"language": ["en", "de", "fr", "es", "pt", "it", "ru", "ko", "nl"], "license": "llama2", "pipeline_tag": "translation"}
translation
BoscoTheDog/TowerInstruct-7B-v0.1-q4f32_1-MLC
[ "translation", "en", "de", "fr", "es", "pt", "it", "ru", "ko", "nl", "license:llama2", "region:us" ]
2024-02-10T15:08:06+00:00
[]
[ "en", "de", "fr", "es", "pt", "it", "ru", "ko", "nl" ]
TAGS #translation #en #de #fr #es #pt #it #ru #ko #nl #license-llama2 #region-us
A WASM version of Tower
[]
[ "TAGS\n#translation #en #de #fr #es #pt #it #ru #ko #nl #license-llama2 #region-us \n" ]
[ 34 ]
[ "passage: TAGS\n#translation #en #de #fr #es #pt #it #ru #ko #nl #license-llama2 #region-us \n" ]
[ 0.06437951326370239, -0.18419931828975677, -0.010122072882950306, 0.03505084291100502, 0.03327684849500656, 0.033867545425891876, 0.07929440587759018, 0.021055109798908234, 0.18238984048366547, -0.0024397114757448435, 0.14380191266536713, 0.04402720183134079, 0.02644507773220539, 0.02284407988190651, 0.053189586848020554, -0.3096376657485962, 0.04657931253314018, -0.051551878452301025, 0.027762582525610924, 0.040894269943237305, 0.07072514295578003, 0.031285472214221954, 0.11134470254182816, -0.025421395897865295, 0.07006366550922394, 0.11300256103277206, -0.02182418666779995, -0.0463290736079216, 0.06483840942382812, 0.08521541208028793, 0.06769657880067825, 0.0018372025806456804, 0.0018866034224629402, -0.24363131821155548, 0.01237198244780302, -0.07541946321725845, -0.15047600865364075, -0.03447193279862404, 0.01995685324072838, -0.12449535727500916, 0.20533455908298492, 0.025346849113702774, -0.0863538384437561, 0.015374339185655117, -0.1818627268075943, -0.1355777233839035, -0.01931520365178585, 0.07607445120811462, 0.042909037321805954, 0.03331737965345383, 0.01135480310767889, 0.0625184178352356, -0.1788201779127121, -0.005554588045924902, 0.14400532841682434, -0.33290672302246094, 0.005498099606484175, 0.12160227447748184, 0.019505782052874565, 0.10454794019460678, -0.037073105573654175, 0.08822120726108551, 0.07019337266683578, -0.04944415017962456, -0.1984466314315796, -0.12589922547340393, 0.08986796438694, 0.05136289820075035, -0.07819172739982605, -0.02694154717028141, 0.1865733414888382, 0.0009300234378315508, 0.01895754598081112, 0.13191361725330353, -0.004753157962113619, -0.08201653510332108, -0.0003286733408458531, -0.05853991582989693, 0.03401181101799011, 0.10863383114337921, 0.2064058631658554, -0.009593825787305832, -0.10363117605447769, 0.008062082342803478, -0.2644452154636383, 0.20567068457603455, 0.045949384570121765, 0.02573772892355919, -0.07607072591781616, -0.028400396928191185, -0.1984090656042099, -0.01510629802942276, -0.08522949367761612, -0.03545410931110382, 0.11973533034324646, 0.04614456743001938, 0.05612456053495407, 0.0508999265730381, 0.10288726538419724, 0.10759327560663223, -0.10528679937124252, 0.019050942733883858, 0.006981454323977232, 0.2062324732542038, 0.03144644573330879, 0.0352025143802166, 0.0455588772892952, -0.006113483104854822, -0.10745653510093689, -0.17751960456371307, 0.01270078681409359, 0.011537938378751278, -0.21088948845863342, -0.02772442251443863, -0.19654373824596405, 0.08471789956092834, -0.09232325106859207, -0.01476797740906477, -0.04809827730059624, 0.05824992433190346, 0.018264587968587875, -0.01573801599442959, 0.008548743091523647, -0.01735941879451275, 0.010787812061607838, 0.13433152437210083, -0.11711962521076202, -0.006565444637089968, 0.06282009929418564, 0.10531648993492126, -0.08219688385725021, 0.06272689998149872, -0.04740486666560173, -0.00741501385346055, 0.0711144432425499, -0.05392715707421303, 0.08769746869802475, -0.18453097343444824, -0.05163022503256798, -0.03323855623602867, -0.006993356626480818, -0.07655937969684601, 0.15861879289150238, -0.06193307787179947, 0.009714215062558651, 0.0451771505177021, -0.06803478300571442, -0.1684819459915161, -0.11468008905649185, 0.05461018159985542, -0.035794783383607864, 0.05238049104809761, -0.18715828657150269, -0.017929593101143837, -0.11671695113182068, 0.08606885373592377, -0.022843388840556145, -0.011274568736553192, -0.13217364251613617, 0.013319558463990688, -0.0471900999546051, -0.018217816948890686, -0.16899985074996948, 0.09128489345312119, -0.06695371866226196, 0.14697293937206268, -0.19940844178199768, -0.07572446018457413, 0.22708207368850708, -0.13045696914196014, -0.049778763204813004, 0.037315912544727325, 0.039194002747535706, 0.09901956468820572, -0.007955980487167835, 0.33905908465385437, -0.05297097936272621, -0.1180097833275795, 0.03484189882874489, 0.19408629834651947, -0.017159242182970047, -0.050007302314043045, 0.11981087923049927, -0.08210857957601547, -0.04091819003224373, 0.012087471783161163, -0.05238838493824005, 0.09908619523048401, -0.03998991847038269, -0.05110463500022888, 0.08255698531866074, 0.03488937392830849, 0.05937004089355469, -0.013753083534538746, 0.05883190408349037, -0.07783223688602448, -0.0001328665966866538, 0.077823705971241, 0.08422034978866577, 0.10920390486717224, -0.0010201934492215514, -0.07362822443246841, 0.04825955256819725, 0.07724263519048691, 0.016041429713368416, 0.008237200789153576, 0.02283952571451664, -0.04923975467681885, 0.03788210451602936, 0.07711489498615265, 0.11995025724172592, 0.09146531671285629, -0.08052634447813034, 0.0023538563400506973, 0.008277365006506443, -0.01725858822464943, 0.0222011748701334, 0.03677480295300484, -0.12338481098413467, 0.1436920464038849, 0.0022198958322405815, -0.12402281910181046, -0.13875572383403778, -0.05059800297021866, 0.17307105660438538, 0.048149384558200836, -0.019466904923319817, 0.10040250420570374, -0.13855120539665222, 0.09269740432500839, -0.028613034635782242, 0.02892531268298626, 0.08820020407438278, -0.09928582608699799, -0.0760243684053421, 0.25228917598724365, -0.03198036551475525, 0.20128829777240753, 0.16586622595787048, -0.11605348438024521, -0.03288877010345459, -0.04919588565826416, -0.01931382715702057, 0.027106868103146553, 0.11907484382390976, -0.024141604080796242, 0.010315489955246449, -0.00011488109157653525, -0.005460706073790789, -0.0651208832859993, 0.021607326343655586, -0.0015814282232895494, -0.12868058681488037, -0.1214042603969574, 0.11834454536437988, 0.18728983402252197, -0.12631158530712128, 0.206454798579216, 0.39595451951026917, 0.08387172222137451, 0.285346120595932, -0.0574333593249321, -0.0015940682496875525, -0.08456707000732422, 0.059350281953811646, -0.03744633123278618, 0.12592852115631104, -0.06589275598526001, -0.016253894194960594, 0.0027084758039563894, 0.06531796604394913, 0.042645104229450226, -0.10180968791246414, -0.12581941485404968, -0.04482869803905487, -0.003420497290790081, -0.04693946614861488, 0.11031976342201233, -0.12123045325279236, 0.0177907757461071, 0.0101450365036726, -0.17681093513965607, 0.03806160390377045, 0.016544319689273834, -0.004525606520473957, 0.12974035739898682, -0.1632019728422165, -0.17525991797447205, -0.09771677851676941, -0.06474202126264572, -0.03616638109087944, 0.001594108180142939, 0.024136623367667198, -0.09421493858098984, -0.033441487699747086, 0.06222088634967804, 0.12156654894351959, -0.16065749526023865, -0.10425872355699539, -0.12231586873531342, 0.05025039613246918, -0.16131097078323364, -0.03179960697889328, -0.09931743890047073, -0.046176549047231674, -0.08473605662584305, 0.08218123763799667, -0.16095800697803497, 0.06983359158039093, 0.10202757269144058, 0.04726501926779747, 0.06138681620359421, -0.06547656655311584, 0.14743027091026306, -0.1580916941165924, -0.06916484236717224, 0.01100520696491003, 0.003312222193926573, 0.04446885734796524, 0.17728568613529205, 0.04557770863175392, -0.06082908809185028, -0.028934910893440247, 0.01205373927950859, -0.10541629791259766, -0.1881566345691681, -0.07021058350801468, -0.07748311758041382, 0.16412027180194855, -0.06166645884513855, 0.05927756056189537, 0.10044486820697784, 0.03281767666339874, 0.12842655181884766, -0.06279502809047699, -0.02309996262192726, -0.0029797854367643595, 0.2257484346628189, -0.07363392412662506, 0.014486134052276611, -0.09491413831710815, -0.03954751044511795, 0.16232843697071075, 0.12245697528123856, 0.0011244835332036018, 0.2128991335630417, 0.005879916716367006, 0.16175635159015656, 0.10895789414644241, 0.05115746706724167, -0.026275096461176872, -0.0026373188011348248, -0.040593601763248444, -0.03049589693546295, -0.039858825504779816, 0.009870664216578007, 0.04594982787966728, 0.0675981193780899, -0.14757759869098663, 0.02779567614197731, -0.09669195115566254, 0.11100225895643234, -0.11636179685592651, 0.07224386185407639, 0.053195010870695114, 0.07559983432292938, 0.06713583320379257, 0.05573159083724022, -0.012842943891882896, 0.17661051452159882, 0.06514853239059448, -0.127802774310112, 0.06388550996780396, 0.10766182094812393, 0.05053120478987694, -0.007831013761460781, 0.04597200080752373, -0.12030106782913208, -0.1131647527217865, 0.027776455506682396, 0.0661802887916565, -0.24805502593517303, 0.28653785586357117, 0.03307498246431351, -0.06418901681900024, 0.030562277883291245, -0.11634252220392227, -0.0018767605070024729, 0.2207980901002884, 0.1599683314561844, 0.1297776699066162, -0.2280263453722, -0.15714478492736816, 0.03279641270637512, -0.026248062029480934, 0.17417441308498383, -0.016254007816314697, -0.08519786596298218, -0.035809699445962906, 0.055521368980407715, -0.06607223302125931, 0.16276255249977112, -0.05780831724405289, -0.08452199399471283, 0.03593241423368454, 0.031054800376296043, 0.06817597150802612, -0.017732111737132072, 0.0005476631340570748, -0.0263367947191, -0.002594010205939412, -0.10634352266788483, 0.032752491533756256, -0.06747112423181534, -0.18988704681396484, 0.04147176817059517, -0.09673313796520233, -0.019236065447330475, -0.02931499108672142, -0.16949696838855743, -0.0963427722454071, -0.022703588008880615, 0.15178804099559784, -0.055201832205057144, 0.024014540016651154, 0.01493912935256958, 0.10676873475313187, 0.003236732678487897, 0.0504356250166893, 0.010079940780997276, -0.00812761764973402, 0.02041776478290558, -0.14038898050785065, 0.07933487743139267, -0.0415249839425087, -0.12938624620437622, 0.002473988104611635, -0.07105132192373276, 0.01843092404305935, -0.0072298129089176655, -0.1091143935918808, 0.1445995718240738, 0.30122873187065125, -0.0056009129621088505, 0.11749399453401566, 0.17430269718170166, -0.09958420693874359, -0.2646651268005371, -0.12331347912549973, -0.18303731083869934, -0.04977795109152794, 0.05906278267502785, -0.1471611112356186, -0.04472920298576355, 0.09180327504873276, -0.022563599050045013, 0.22527222335338593, -0.31175991892814636, -0.053292166441679, 0.11299824714660645, -0.06893650442361832, 0.44701993465423584, -0.047669071704149246, -0.10760156810283661, 0.0007394528947770596, -0.154573455452919, 0.08406778424978256, 0.027181515470147133, 0.05060181766748428, 0.015478168614208698, 0.055912695825099945, -0.014924533665180206, 0.05960867926478386, 0.2000918686389923, 0.026322565972805023, 0.02863547019660473, -0.08724390715360641, -0.17500415444374084, 0.016305528581142426, 0.045362044125795364, -0.05204429104924202, -0.02323155850172043, -0.07388488948345184, -0.16226594150066376, -0.009935143403708935, -0.06077546998858452, 0.10055646300315857, -0.002104024402797222, -0.08876853436231613, -0.08321883529424667, 0.08636897802352905, -0.06285014748573303, 0.018604004755616188, 0.31566882133483887, -0.05095149576663971, 0.04041265323758125, 0.016420455649495125, -0.01731327548623085, 0.02905178628861904, 0.06219431757926941, -0.06596121191978455, -0.05310772731900215, 0.07231301069259644, -0.08892957121133804, -0.060225874185562134, 0.1242249608039856, -0.06535863876342773, 0.03704071044921875, 0.013635527342557907, -0.1249452605843544, 0.05139702185988426, 0.16111592948436737, -0.030480939894914627, -0.17909763753414154, -0.02947819232940674, -0.07399863004684448, 0.2735205590724945, -0.04719064012169838, 0.11527219414710999, -0.0009527222136966884, 0.018465155735611916, 0.03305220603942871, -0.012833076529204845, -0.12207435071468353, -0.027072975412011147, 0.0280739888548851, -0.053421974182128906, -0.04414753615856171, 0.09343674033880234, -0.014183254912495613, -0.13006055355072021, 0.0057383328676223755, 0.11579423397779465, -0.0614326111972332, -0.09318742901086807, -0.132720947265625, 0.06404739618301392, -0.2251332700252533, -0.0942651778459549, 0.06869182735681534, -0.09780009835958481, -0.013201561756432056, 0.17156049609184265, 0.021818283945322037, 0.040544480085372925, 0.048268284648656845, 0.005073007196187973, 0.21453940868377686, -0.01321338675916195, -0.06465879082679749, -0.019260600209236145, -0.02844896726310253, -0.05448854714632034, 0.02734198048710823, 0.1391717493534088, -0.0397089384496212, -0.06317105144262314, -0.22099585831165314, 0.021819772198796272, -0.03882116451859474, -0.03998547047376633, -0.10330236703157425, -0.03828667476773262, 0.047969233244657516, -0.10144326835870743, -0.08809637278318405, -0.06810221821069717, -0.17668728530406952, 0.038788486272096634, 0.0474899522960186, 0.1314515769481659, -0.015517530962824821, -0.014192324131727219, 0.11401667445898056, 0.04211019352078438, 0.03164391219615936, 0.12906695902347565, 0.08450784534215927, 0.2175683230161667, -0.15615859627723694, 0.08990377187728882, 0.08155577629804611, 0.013844533823430538, 0.07341776043176651, 0.18712696433067322, -0.04885636642575264, 0.023825347423553467, 0.014908485114574432, 0.03757715970277786, -0.10545448213815689, -0.08679372817277908, 0.02865445241332054, -0.023418201133608818, -0.19452355802059174, 0.045411039143800735, -0.026597732678055763, 0.08039242774248123, -0.023641187697649002, 0.09956236928701401, -0.022548982873558998, 0.05304988473653793, 0.11525961011648178, 0.024020418524742126, 0.00293768965639174, -0.09605152904987335, 0.005482947453856468, -0.12891940772533417, -0.05953517183661461, -0.004390494432300329, 0.24180863797664642, 0.12983191013336182, -0.0353146530687809, 0.006465947721153498, -0.0024686965625733137, -0.07628211379051208, 0.004797924309968948, 0.10008537024259567, 0.1033172532916069, -0.04565756022930145, -0.2735999822616577, 0.03291338309645653, -0.05849860981106758, -0.1658180058002472, 0.04357688128948212, 0.08871752768754959, 0.0706641674041748, 0.08519332110881805, 0.036439042538404465, -0.018269384279847145, 0.01260787807404995, -0.1667051464319229, 0.059260983020067215, -0.046127043664455414, -0.031090281903743744, 0.1259382963180542, 0.21379415690898895, -0.0600527822971344, 0.023096397519111633, -0.02092999778687954, 0.007450662087649107, -0.18322396278381348, -0.10907385498285294, 0.01530646812170744, -0.1674332171678543, 0.04206414520740509, -0.001453072763979435, 0.07388538122177124, 0.041014786809682846, 0.10122599452733994, -0.019791370257735252, 0.09416386485099792, -0.15643370151519775, -0.1209271177649498, 0.019871536642313004, -0.01980421133339405, 0.10448303073644638, -0.18140754103660583, -0.05697719752788544, -0.12690739333629608, -0.1388465017080307, -0.07646838575601578, 0.04406581446528435, -0.004172537010163069, -0.08052418380975723, -0.12958142161369324, -0.018401024863123894, -0.02462988719344139, 0.10081031173467636, -0.059073399752378464, 0.11908545345067978, 0.03329547867178917, -0.009802316315472126, 0.011451502330601215, 0.09080713987350464, 0.042542826384305954, -0.037139132618904114, 0.042230915278196335, 0.08157306909561157, 0.0463879220187664, 0.21464002132415771, -0.10560479015111923, 0.015389809384942055, -0.060002345591783524, 0.176140695810318, 0.24122227728366852, -0.04303717613220215, 0.030281255021691322, 0.05158112943172455, 0.08369550853967667, 0.08498649299144745, 0.05779927596449852, -0.013252723962068558, 0.2806020677089691, -0.05394342541694641, 0.006751516833901405, -0.09538313001394272, 0.04460165277123451, -0.03258528560400009, 0.028067896142601967, 0.07990724593400955, -0.042614132165908813, -0.07587332278490067, 0.13544347882270813, -0.2561820149421692, 0.04722204804420471, 0.10337831825017929, -0.1339222937822342, 0.0034215913619846106, -0.0436561219394207, 0.09447051584720612, 0.12687897682189941, 0.10496021807193756, -0.05917331948876381, -0.17540690302848816, -0.08170435577630997, 0.0982552170753479, -0.2724551260471344, -0.1409471482038498, 0.02807631529867649, 0.13308605551719666, 0.12725506722927094, -0.022935383021831512, -0.014086193405091763, 0.03671819344162941, 0.04285368695855141, 0.06221642717719078, 0.07594694942235947, 0.07647190243005753, 0.1023726537823677, -0.17070119082927704, -0.1660945862531662, 0.037778545171022415, -0.08243588358163834, 0.10538286715745926, -0.04171857610344887, 0.06863721460103989, 0.1180792823433876, -0.09467267990112305, -0.02781011536717415, 0.1258183717727661, -0.14068472385406494, -0.00438727205619216, 0.058249205350875854, 0.06539328396320343, -0.048405591398477554, -0.054671794176101685, 0.0038387952372431755, 0.06100805103778839, -0.10747817903757095, -0.04635581001639366, 0.16238175332546234, -0.04474512115120888, 0.08972674608230591, 0.041178036481142044, -0.05848408490419388, 0.026214836165308952, -0.04164225980639458, 0.1575135737657547, -0.092599056661129, 0.0758446604013443, 0.0738348737359047, -0.014000920578837395, 0.03779098019003868, -0.3070163130760193, 0.08203345537185669, -0.023409252986311913, 0.032103702425956726, -0.002987202489748597 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # trainer This model is a fine-tuned version of [bert-large-uncased](https://huggingface.co/bert-large-uncased) on the glue dataset. It achieves the following results on the evaluation set: - Loss: 0.4490 - Accuracy: 0.8466 - F1: 0.8065 - Precision: 0.8406 - Recall: 0.7876 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 8 - eval_batch_size: 32 - seed: 42 - gradient_accumulation_steps: 4 - total_train_batch_size: 32 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | Precision | Recall | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:|:---------:|:------:| | No log | 1.0 | 267 | 0.3860 | 0.8370 | 0.7999 | 0.8184 | 0.7876 | | 0.3455 | 2.0 | 534 | 0.4490 | 0.8466 | 0.8065 | 0.8406 | 0.7876 | ### Framework versions - Transformers 4.37.0 - Pytorch 2.1.2 - Datasets 2.1.0 - Tokenizers 0.15.1
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "datasets": ["glue"], "metrics": ["accuracy", "f1", "precision", "recall"], "base_model": "bert-large-uncased", "model-index": [{"name": "trainer", "results": [{"task": {"type": "text-classification", "name": "Text Classification"}, "dataset": {"name": "glue", "type": "glue", "config": "cola", "split": "validation", "args": "cola"}, "metrics": [{"type": "accuracy", "value": 0.8465963566634708, "name": "Accuracy"}, {"type": "f1", "value": 0.8064540073113251, "name": "F1"}, {"type": "precision", "value": 0.840606542828289, "name": "Precision"}, {"type": "recall", "value": 0.7876439727431708, "name": "Recall"}]}]}]}
text-classification
Sifal/bertGED
[ "transformers", "tensorboard", "safetensors", "bert", "text-classification", "generated_from_trainer", "dataset:glue", "base_model:bert-large-uncased", "license:apache-2.0", "model-index", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-10T15:09:33+00:00
[]
[]
TAGS #transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #dataset-glue #base_model-bert-large-uncased #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us
trainer ======= This model is a fine-tuned version of bert-large-uncased on the glue dataset. It achieves the following results on the evaluation set: * Loss: 0.4490 * Accuracy: 0.8466 * F1: 0.8065 * Precision: 0.8406 * Recall: 0.7876 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 8 * eval\_batch\_size: 32 * seed: 42 * gradient\_accumulation\_steps: 4 * total\_train\_batch\_size: 32 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 2 ### Training results ### Framework versions * Transformers 4.37.0 * Pytorch 2.1.2 * Datasets 2.1.0 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.1.0\n* Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #dataset-glue #base_model-bert-large-uncased #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.1.0\n* Tokenizers 0.15.1" ]
[ 79, 126, 4, 30 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #bert #text-classification #generated_from_trainer #dataset-glue #base_model-bert-large-uncased #license-apache-2.0 #model-index #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 32\n* seed: 42\n* gradient\\_accumulation\\_steps: 4\n* total\\_train\\_batch\\_size: 32\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2### Training results### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.1.0\n* Tokenizers 0.15.1" ]
[ -0.12863224744796753, 0.15584468841552734, -0.0036019557155668736, 0.09342149645090103, 0.12807786464691162, 0.021281085908412933, 0.13220587372779846, 0.15392638742923737, -0.08490537852048874, 0.057782094925642014, 0.11011139303445816, 0.09977945685386658, 0.056560296565294266, 0.16635672748088837, -0.029808780178427696, -0.2922585606575012, 0.005831948947161436, 0.032395027577877045, -0.12018568813800812, 0.140341654419899, 0.09935013204813004, -0.11504309624433517, 0.06061159446835518, 0.01519207376986742, -0.12916980683803558, -0.022464171051979065, 0.0049255588091909885, -0.07703286409378052, 0.11513787508010864, 0.0024761727545410395, 0.11988998204469681, 0.062328681349754333, 0.09222053736448288, -0.1602667272090912, 0.009861839935183525, 0.07213732600212097, 0.014224744401872158, 0.11390829086303711, 0.08404640853404999, -0.013065735809504986, 0.09961266070604324, -0.09034869074821472, 0.06972383707761765, 0.036699239164590836, -0.10012905299663544, -0.3155818283557892, -0.09198649227619171, 0.09470370411872864, 0.10882923007011414, 0.06091860681772232, -0.0011436776258051395, 0.086228147149086, -0.06107565388083458, 0.07607635110616684, 0.26203233003616333, -0.31921204924583435, -0.05695229768753052, 0.019206907600164413, 0.027931392192840576, 0.04731695353984833, -0.1047644391655922, -0.02618207037448883, 0.026719488203525543, 0.03338473290205002, 0.15283389389514923, 0.0005808091373182833, 0.03649338334798813, -0.005598674528300762, -0.1459616720676422, -0.058263324201107025, 0.12485425174236298, 0.038006387650966644, -0.041410692036151886, -0.09221470355987549, -0.0705140009522438, -0.16582590341567993, -0.020930586382746696, -0.004389762878417969, 0.027332298457622528, -0.05211721360683441, -0.09382010251283646, -0.02824312075972557, -0.09424126148223877, -0.07335459440946579, 0.0003978772147092968, 0.12515757977962494, 0.06575173139572144, 0.023205000907182693, 0.004489895887672901, 0.14931011199951172, 0.05722900107502937, -0.1595500409603119, 0.015324183739721775, 0.005948987789452076, -0.026760540902614594, -0.0032454300671815872, -0.018023164942860603, -0.031364019960165024, -0.0016640248941257596, 0.12873011827468872, -0.09148634225130081, 0.021359190344810486, 0.02485511265695095, 0.05124405398964882, -0.09430862963199615, 0.16988518834114075, -0.0692567452788353, -0.04472597688436508, 0.003126545110717416, 0.1312437802553177, 0.05949757620692253, -0.02203848399221897, -0.10773894935846329, 0.00851005595177412, 0.1335296630859375, 0.05924195423722267, 0.0050809551030397415, 0.030793163925409317, -0.05758697912096977, -0.05627167224884033, 0.08329755067825317, -0.10501540452241898, 0.028575005009770393, 0.02500898577272892, -0.08387012034654617, -0.02916610799729824, 0.02231748029589653, -0.006082508247345686, -0.030297987163066864, 0.11473046988248825, -0.07640714198350906, -0.0019871455151587725, -0.08199073374271393, -0.09858685731887817, 0.02366742677986622, -0.07164809852838516, -0.018548866733908653, -0.08688021451234818, -0.16855889558792114, -0.010833757929503918, 0.05538737028837204, -0.048826269805431366, -0.08289645612239838, -0.021562770009040833, -0.08185145258903503, 0.030568178743124008, -0.03362967446446419, 0.09618870913982391, -0.0414259247481823, 0.13229374587535858, 0.020944003015756607, 0.061555165797472, 0.03879842534661293, 0.047946128994226456, -0.09000908583402634, 0.052029699087142944, -0.1804346889257431, 0.04399746656417847, -0.05525951832532883, 0.015848664566874504, -0.11524272710084915, -0.13132090866565704, 0.03110087476670742, -0.03520617634057999, 0.07785290479660034, 0.10226812213659286, -0.1384577751159668, -0.09215264022350311, 0.17327845096588135, -0.11869147419929504, -0.11407127976417542, 0.12672114372253418, -0.007068044040352106, -0.01602564938366413, 0.03684493526816368, 0.15606452524662018, 0.09385818988084793, -0.07431686669588089, -0.018072599545121193, -0.0037478397134691477, 0.08019337803125381, -0.03880464285612106, 0.13030306994915009, 0.00027476888499222696, 0.017228737473487854, 0.020971322432160378, -0.07963519543409348, 0.0521802194416523, -0.09560797363519669, -0.09282252192497253, -0.027383271604776382, -0.07869880646467209, 0.11716769635677338, 0.07622136175632477, 0.058963529765605927, -0.08121536672115326, -0.13266456127166748, 0.027786454185843468, 0.1113743856549263, -0.08199042081832886, 0.030293969437479973, -0.0664796456694603, 0.12848304212093353, -0.07925911247730255, -0.01474571879953146, -0.17173680663108826, -0.054674457758665085, 0.031556662172079086, -0.06009523198008537, -0.022595452144742012, -0.0034738953690975904, 0.06273239105939865, 0.09440246969461441, -0.08595303446054459, -0.08100332319736481, -0.07298162579536438, -0.011222798377275467, -0.09180935472249985, -0.22144021093845367, -0.09306284040212631, -0.025830712169408798, 0.1304064393043518, -0.18017415702342987, 0.050564974546432495, 0.002947190310806036, 0.13252972066402435, 0.04194621369242668, -0.026705186814069748, -0.0142886433750391, 0.05634566769003868, -0.03115667775273323, -0.07115772366523743, 0.0755278691649437, 0.0189796332269907, -0.09324818104505539, 0.002369569381698966, -0.11659839004278183, 0.1681952327489853, 0.1095738410949707, 0.0019208767917007208, -0.0682254210114479, -0.025154776871204376, -0.09406563639640808, -0.03500562161207199, -0.04386146366596222, 0.011275168508291245, 0.10064271837472916, 0.022500256076455116, 0.16255968809127808, -0.09981907159090042, -0.049041636288166046, 0.03096296824514866, -0.014586430042982101, -0.003966918680816889, 0.11575639247894287, 0.05120813474059105, -0.06154442951083183, 0.1387677937746048, 0.1741493046283722, -0.058450847864151, 0.13014675676822662, -0.05809644237160683, -0.08814406394958496, -0.04559823125600815, -0.004104215186089277, 0.02144152857363224, 0.13235150277614594, -0.08246783167123795, 0.0051828790456056595, 0.04095081612467766, 0.019762884825468063, 0.012862573377788067, -0.20571960508823395, -0.012902437709271908, 0.02966875210404396, -0.06606226414442062, -0.007922765798866749, -0.009107117541134357, 0.016905618831515312, 0.1152220070362091, 0.00752642285078764, -0.057463403791189194, 0.03465806320309639, 0.00016848430095706135, -0.07806719839572906, 0.19296902418136597, -0.07064977288246155, -0.1962614357471466, -0.1371898055076599, 0.012045607902109623, -0.0582394078373909, -0.008019139990210533, 0.0488630011677742, -0.06570978462696075, -0.015245655551552773, -0.07591455429792404, 0.03389701619744301, -0.009319021366536617, 0.041335344314575195, 0.010774267837405205, 0.02153671719133854, 0.08993767946958542, -0.1016089916229248, 0.009491778910160065, -0.03364979103207588, -0.060897599905729294, 0.03547303378582001, 0.06271238625049591, 0.08662974834442139, 0.15440860390663147, -0.011259786784648895, 0.023332322016358376, -0.03994739428162575, 0.145906463265419, -0.07013767212629318, -0.006514717359095812, 0.16898328065872192, -0.0029616612009704113, 0.07333627343177795, 0.13432185351848602, 0.05467609316110611, -0.037155862897634506, -0.012598400935530663, 0.021476536989212036, -0.03455773741006851, -0.2318788319826126, -0.04650723561644554, -0.03423922136425972, 0.04706490412354469, 0.11744245141744614, 0.032290562987327576, 0.009073031134903431, 0.05018456652760506, -0.02649237960577011, 0.016200846061110497, -0.004131177440285683, 0.07410826534032822, 0.07785744220018387, 0.029636232182383537, 0.1371358036994934, -0.030118275433778763, -0.030425339937210083, 0.03830835968255997, -0.0011257035657763481, 0.2082272320985794, -0.0795295313000679, 0.12142499536275864, 0.05054323747754097, 0.18263976275920868, 0.02045898325741291, 0.07706817984580994, -0.004889432340860367, -0.0038848069962114096, 0.00021219921472948045, -0.04942130669951439, -0.031856387853622437, 0.014152544550597668, 0.001863181358203292, 0.08009827136993408, -0.15441793203353882, 0.009019782766699791, 0.044613249599933624, 0.2895524799823761, 0.09646224975585938, -0.3138284683227539, -0.1158081442117691, 0.008601400069892406, -0.024302521720528603, -0.02976381778717041, 0.01245330274105072, 0.12985770404338837, -0.10899369418621063, 0.051896099001169205, -0.10531435161828995, 0.06768051534891129, -0.07391136139631271, 0.005293505731970072, 0.08029567450284958, 0.09377990663051605, 0.016971660777926445, 0.08724328875541687, -0.23074659705162048, 0.2678295969963074, 0.0010988423600792885, 0.05109909921884537, -0.051525987684726715, 0.02201993390917778, 0.02910017780959606, 0.04697473347187042, 0.08974209427833557, -0.005667647812515497, -0.03344777598977089, -0.19249500334262848, -0.1342325657606125, -0.0008116699755191803, 0.08681254833936691, -0.08373499661684036, 0.10147547721862793, -0.03849995508790016, -0.02179161086678505, 0.03500068560242653, -0.053158435970544815, -0.06043402850627899, -0.09646010398864746, 0.03866636008024216, -0.01986733078956604, -0.0006402842700481415, -0.1093859076499939, -0.14089737832546234, -0.05673588439822197, 0.1985970139503479, -0.07805481553077698, -0.08550943434238434, -0.12752071022987366, 0.119669109582901, 0.11444240808486938, -0.09575140476226807, 0.02453739009797573, -0.01033253688365221, 0.13275717198848724, 0.01171961147338152, -0.07020261883735657, 0.07894238829612732, -0.07706128060817719, -0.2500765919685364, -0.054833970963954926, 0.168076753616333, 0.011802499182522297, 0.07269175350666046, -0.021480802446603775, 0.013601983897387981, 0.004031005315482616, -0.07461833208799362, 0.041590381413698196, 0.043659619987010956, 0.07226797193288803, 0.0285023245960474, -0.04040854051709175, 0.006163662765175104, -0.042637716978788376, -0.018608037382364273, 0.15304480493068695, 0.24841414391994476, -0.11502130329608917, 0.07334616780281067, 0.06650272756814957, -0.042805153876543045, -0.19889941811561584, -0.020383300259709358, 0.10269267857074738, 0.014882771298289299, 0.02863623946905136, -0.16361850500106812, 0.05538616701960564, 0.08121410012245178, -0.04878731444478035, 0.08438660949468613, -0.3335115909576416, -0.12016469240188599, 0.07933980971574783, 0.11773260682821274, 0.05765395611524582, -0.166951984167099, -0.06341083347797394, 0.020070211961865425, -0.07976540923118591, 0.10472109913825989, -0.06593229621648788, 0.10957196354866028, -0.04607333242893219, -0.0000383349324692972, 0.016719872131943703, -0.06834615767002106, 0.14019989967346191, -0.03925669193267822, 0.041863080114126205, -0.03153662011027336, 0.006349630653858185, 0.1039285808801651, -0.08354286104440689, 0.032456234097480774, -0.06470577418804169, 0.056371934711933136, -0.09723544120788574, -0.0195348858833313, -0.10764583200216293, 0.038976237177848816, -0.052366308867931366, -0.059175148606300354, -0.025326021015644073, 0.0551467165350914, 0.053665146231651306, -0.020625043660402298, 0.1439359188079834, 0.034491509199142456, 0.1673765927553177, 0.15436653792858124, 0.05426972731947899, -0.02722311206161976, -0.10405154526233673, -0.047883473336696625, -0.02991504780948162, 0.054562654346227646, -0.1427013874053955, 0.03171289339661598, 0.13562363386154175, 0.03365013375878334, 0.12755246460437775, 0.07218064367771149, -0.05633671581745148, 0.023170962929725647, 0.07324447482824326, -0.14391641318798065, -0.1097332164645195, -0.03803909197449684, -0.0042554778046905994, -0.17039504647254944, 0.06168011575937271, 0.1140749379992485, -0.06264886260032654, -0.025262420997023582, 0.012017559260129929, 0.02973785810172558, -0.0062415217980742455, 0.18732981383800507, 0.06865537166595459, 0.08424491435289383, -0.09634292125701904, 0.09449291974306107, 0.06313448399305344, -0.12808698415756226, 0.0030053313821554184, 0.07230806350708008, -0.08612357825040817, -0.02882908284664154, 0.019291458651423454, 0.09501928091049194, -0.023210883140563965, -0.05573021247982979, -0.16360412538051605, -0.11970237642526627, 0.07331416755914688, 0.07125410437583923, 0.0811384916305542, 0.034200605005025864, -0.032564349472522736, 0.015067937783896923, -0.11289232224225998, 0.14696763455867767, 0.07826896756887436, 0.07636427879333496, -0.17469674348831177, 0.1457943618297577, 0.02603014186024666, 0.07096145302057266, -0.01855449192225933, 0.0164823979139328, -0.08375315368175507, -0.01581113412976265, -0.1054464727640152, -0.018281664699316025, -0.05105092376470566, -0.01315407082438469, -0.017979903146624565, -0.05429007112979889, -0.03647680580615997, 0.028576238080859184, -0.10158931463956833, -0.043669234961271286, -0.003959042485803366, 0.03473393991589546, -0.1302250176668167, -0.03137505054473877, 0.0204972792416811, -0.10241826623678207, 0.1086883693933487, 0.023992232978343964, 0.03783707693219185, 0.024457048624753952, -0.05325097218155861, 0.006005029194056988, 0.03049095906317234, -0.0014852422755211592, 0.04303937405347824, -0.12759386003017426, 0.006334479432553053, -0.01600719802081585, -0.01433588471263647, 0.01469823345541954, 0.07211817055940628, -0.1453542411327362, 0.016646699979901314, -0.01967071369290352, -0.043088193982839584, -0.062184203416109085, 0.059128403663635254, 0.047524578869342804, 0.035149965435266495, 0.1860174685716629, -0.08259966224431992, 0.059350352734327316, -0.2381889373064041, 0.0006902696914039552, -0.02669479511678219, -0.08167022466659546, -0.07940815389156342, -0.013594732619822025, 0.06220279633998871, -0.051580626517534256, 0.08315302431583405, 0.013635590672492981, 0.07976274937391281, 0.03106553852558136, -0.052451666444540024, 0.05056559294462204, 0.03661030903458595, 0.1595558524131775, 0.010108107700943947, -0.028168486431241035, 0.06852460652589798, 0.015145291574299335, 0.06793341785669327, 0.09010262042284012, 0.17895467579364777, 0.11976626515388489, 0.02404051460325718, 0.07270827144384384, 0.0353582464158535, -0.06584089249372482, -0.2336256206035614, 0.04558415338397026, -0.029975702986121178, 0.11952007561922073, -0.0005215842393226922, 0.18405607342720032, 0.11755549907684326, -0.19016677141189575, 0.038789309561252594, -0.011780975386500359, -0.07432540506124496, -0.10698508471250534, -0.0814250186085701, -0.07258002460002899, -0.16283057630062103, -0.0065713790245354176, -0.10917486995458603, 0.0028180901426821947, 0.0865965411067009, 0.007831873372197151, 0.008085628971457481, 0.11202136427164078, 0.062352314591407776, 0.024054214358329773, 0.06509626656770706, 0.04045022279024124, -0.003645265242084861, -0.01474206056445837, -0.07392454147338867, 0.00016058297478593886, 0.004713285714387894, 0.043728962540626526, -0.0513039231300354, -0.06158416345715523, 0.04825171083211899, 0.007519331760704517, -0.11118737608194351, 0.015072982758283615, -0.002428955864161253, 0.07649053633213043, 0.07729125022888184, 0.011219083331525326, 0.009371704421937466, -0.03119460493326187, 0.2531030476093292, -0.08827407658100128, -0.04514619708061218, -0.12459473311901093, 0.28713324666023254, 0.014215188100934029, -0.04960409551858902, 0.060344476252794266, -0.09571833908557892, -0.02088429406285286, 0.16653363406658173, 0.20425176620483398, -0.02641991712152958, -0.011573366820812225, 0.020502831786870956, -0.02028689906001091, -0.02504151314496994, 0.09394770860671997, 0.10726483166217804, 0.10518068075180054, -0.09164255112409592, -0.030902806669473648, -0.05014437809586525, -0.01938943937420845, -0.04327087476849556, 0.09950175881385803, 0.008692612871527672, 0.006478567607700825, -0.048137519508600235, 0.06039416417479515, -0.006610856857150793, -0.12136863172054291, 0.06869714707136154, -0.21517173945903778, -0.18493011593818665, -0.017988603562116623, 0.10686160624027252, -0.0008362612570635974, 0.06131872162222862, 0.000970824621617794, 0.0061986371874809265, 0.1009998545050621, -0.015126888640224934, -0.06954628974199295, -0.0735696330666542, 0.09602565318346024, -0.0930749848484993, 0.2330539971590042, -0.040029849857091904, 0.028648922219872475, 0.1218990758061409, 0.03600539267063141, -0.11877044290304184, 0.019717393442988396, 0.0630163624882698, -0.06313261389732361, 0.03130828216671944, 0.1678829938173294, -0.016160348430275917, 0.07990331947803497, 0.04370926693081856, -0.1267285943031311, -0.014907635748386383, -0.06676662713289261, -0.07333998382091522, -0.043522439897060394, -0.0007567977881990373, -0.02090069092810154, 0.12245649844408035, 0.205558180809021, -0.05614306777715683, -0.014835941605269909, -0.06526795029640198, 0.006840612273663282, 0.06051712855696678, 0.066570945084095, 0.009379023686051369, -0.2266434282064438, 0.023570654913783073, 0.020077068358659744, 0.01866758055984974, -0.26449844241142273, -0.09475678950548172, 0.031222499907016754, -0.06873606890439987, -0.12154422700405121, 0.09630636870861053, 0.028853505849838257, 0.047416605055332184, -0.03993651643395424, -0.002422262914478779, -0.07980051636695862, 0.1498037576675415, -0.1879713535308838, -0.06953710317611694 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # GeneralNews_9_loadbest This model is a fine-tuned version of [csebuetnlp/mT5_multilingual_XLSum](https://huggingface.co/csebuetnlp/mT5_multilingual_XLSum) on an unknown dataset. It achieves the following results on the evaluation set: - Loss: 2.7052 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 1000 - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 3.4665 | 1.0 | 500 | 3.0865 | | 3.2219 | 2.0 | 1000 | 2.8680 | | 3.0254 | 3.0 | 1500 | 2.7771 | | 2.8798 | 4.0 | 2000 | 2.7572 | | 2.7943 | 5.0 | 2500 | 2.7339 | | 2.6988 | 6.0 | 3000 | 2.7126 | | 2.4587 | 7.0 | 3500 | 2.7086 | | 2.4769 | 8.0 | 4000 | 2.7076 | | 2.5388 | 9.0 | 4500 | 2.7075 | | 2.4411 | 10.0 | 5000 | 2.7052 | ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.1
{"tags": ["generated_from_trainer"], "base_model": "csebuetnlp/mT5_multilingual_XLSum", "model-index": [{"name": "GeneralNews_9_loadbest", "results": []}]}
text2text-generation
tjkmitl/GeneralNews_2_loadbest_10000rows
[ "transformers", "tensorboard", "safetensors", "mt5", "text2text-generation", "generated_from_trainer", "base_model:csebuetnlp/mT5_multilingual_XLSum", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T15:11:03+00:00
[]
[]
TAGS #transformers #tensorboard #safetensors #mt5 #text2text-generation #generated_from_trainer #base_model-csebuetnlp/mT5_multilingual_XLSum #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
GeneralNews\_9\_loadbest ======================== This model is a fine-tuned version of csebuetnlp/mT5\_multilingual\_XLSum on an unknown dataset. It achieves the following results on the evaluation set: * Loss: 2.7052 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 5e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 1000 * num\_epochs: 10 ### Training results ### Framework versions * Transformers 4.35.2 * Pytorch 2.1.0+cu121 * Datasets 2.17.0 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10", "### Training results", "### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #tensorboard #safetensors #mt5 #text2text-generation #generated_from_trainer #base_model-csebuetnlp/mT5_multilingual_XLSum #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10", "### Training results", "### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ 83, 116, 4, 33 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #mt5 #text2text-generation #generated_from_trainer #base_model-csebuetnlp/mT5_multilingual_XLSum #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 1000\n* num\\_epochs: 10### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ -0.1456754207611084, 0.13546152412891388, -0.0016651819460093975, 0.09868098795413971, 0.12807321548461914, 0.023050149902701378, 0.16873082518577576, 0.12122264504432678, -0.042738672345876694, 0.09066737443208694, 0.15380936861038208, 0.10357966274023056, 0.0408659465610981, 0.18279515206813812, -0.06991938501596451, -0.21315975487232208, 0.03006027266383171, 0.01267997082322836, -0.07147017121315002, 0.1296820193529129, 0.07936185598373413, -0.12388917058706284, 0.09414272010326385, -0.015664605423808098, -0.16213944554328918, -0.042164113372564316, -0.001962884096428752, -0.044284380972385406, 0.12185563892126083, 0.02811782993376255, 0.1108970195055008, 0.05666887015104294, 0.07650389522314072, -0.1729440838098526, 0.010126524604856968, 0.04264787212014198, 0.0020979929249733686, 0.09615348279476166, 0.0545758455991745, -0.01653105393052101, 0.05003281682729721, -0.10176599770784378, 0.04676024243235588, 0.007815900258719921, -0.12444013357162476, -0.1408402919769287, -0.09219648689031601, 0.044575005769729614, 0.08492299169301987, 0.07137753814458847, -0.010777726769447327, 0.12998297810554504, -0.03275183215737343, 0.10167234390974045, 0.21725556254386902, -0.30236658453941345, -0.06794914603233337, 0.028644952923059464, 0.05584678053855896, 0.10453888773918152, -0.0974430963397026, -0.011846550740301609, 0.04880761355161667, 0.01538615021854639, 0.1191457062959671, -0.011670570820569992, 0.010480563156306744, -0.008224241435527802, -0.1419462263584137, -0.050848137587308884, 0.16019359230995178, 0.0458308644592762, -0.04500379413366318, -0.08154643326997757, -0.07495088130235672, -0.17902924120426178, -0.02493678592145443, 0.0026961504481732845, 0.01361167710274458, -0.04262879490852356, -0.08783682435750961, -0.02990521863102913, -0.09297949075698853, -0.07398561388254166, -0.003232583636417985, 0.14268438518047333, 0.02740289457142353, 0.004913114011287689, -0.005561549216508865, 0.10303953289985657, 0.017592526972293854, -0.16045860946178436, 0.0074484641663730145, 0.004534546285867691, -0.03351433575153351, -0.033881645649671555, -0.04854898527264595, -0.0604032427072525, 0.014888036996126175, 0.1388942450284958, -0.019236428663134575, 0.05536552891135216, 0.01838887669146061, 0.025423085317015648, -0.07921135425567627, 0.1529262810945511, -0.05524993687868118, -0.06545856595039368, 0.012621124275028706, 0.12893925607204437, 0.04729476571083069, -0.036421649158000946, -0.10631722211837769, 0.000507858581840992, 0.1333494335412979, 0.03619822859764099, -0.02692526765167713, 0.0529300719499588, -0.05680456757545471, -0.026568826287984848, 0.06763191521167755, -0.09981238096952438, 0.016138330101966858, 0.012521950528025627, -0.051536060869693756, -0.04754497483372688, 0.010929415933787823, 0.013586780056357384, -0.0008004511473700404, 0.07978881895542145, -0.10348748415708542, -0.002838388318195939, -0.06814327090978622, -0.12321438640356064, 0.016425609588623047, -0.1042989119887352, -0.0049018459394574165, -0.08963894098997116, -0.1738465428352356, -0.03410456329584122, 0.030320748686790466, -0.041586294770240784, -0.05890018865466118, -0.07570751011371613, -0.09129755198955536, 0.04095148667693138, -0.010167251341044903, 0.08881653100252151, -0.05933355167508125, 0.10194087028503418, 0.03680136799812317, 0.07462529093027115, -0.027685614302754402, 0.03364654630422592, -0.07259664684534073, 0.05941244959831238, -0.14697235822677612, 0.0739852711558342, -0.05665555223822594, 0.049954384565353394, -0.0906599834561348, -0.08307824283838272, -0.002956975484266877, -0.0211808905005455, 0.09017150104045868, 0.13514690101146698, -0.16677336394786835, -0.059999044984579086, 0.19525165855884552, -0.08778094500303268, -0.15196284651756287, 0.13134276866912842, -0.04042144864797592, 0.004614613950252533, 0.050466328859329224, 0.15964867174625397, 0.09808803349733353, -0.08904734253883362, -0.02820505015552044, -0.0166982002556324, 0.0908362865447998, -0.048622533679008484, 0.11136484891176224, 0.01733858324587345, 0.011606364510953426, 0.0014747194945812225, -0.06301382929086685, 0.04784966632723808, -0.09216837584972382, -0.09419633448123932, -0.033072579652071, -0.10268235951662064, 0.06495939940214157, 0.04255705699324608, 0.058979518711566925, -0.09742935001850128, -0.08698708564043045, 0.031178375706076622, 0.0934973806142807, -0.0970175638794899, 0.00954768992960453, -0.06655637919902802, 0.0955936461687088, -0.10796960443258286, -0.020940439775586128, -0.1696869134902954, -0.08815163373947144, 0.0289604589343071, 0.008841821923851967, -0.007458292413502932, -0.019589530304074287, 0.07944190502166748, 0.09960990399122238, -0.04980665445327759, -0.057529594749212265, -0.02757345698773861, -0.005464536137878895, -0.12286777049303055, -0.19068285822868347, -0.057267554104328156, -0.04440346360206604, 0.20528024435043335, -0.22670145332813263, 0.03480644151568413, 0.0054473672062158585, 0.10353484749794006, 0.038802675902843475, -0.04020265117287636, -0.005775678437203169, 0.06568344682455063, -0.05025319382548332, -0.07258576899766922, 0.056198131293058395, 0.019123150035738945, -0.12122707068920135, 0.008926894515752792, -0.16639360785484314, 0.14849485456943512, 0.12124568969011307, 0.009917289018630981, -0.06741216033697128, -0.03163415566086769, -0.050602421164512634, -0.041598640382289886, -0.02631690725684166, -0.018102183938026428, 0.10159298777580261, 0.019913116469979286, 0.14242799580097198, -0.08938650786876678, -0.04694288596510887, 0.0303779486566782, -0.01249726489186287, -0.014175611548125744, 0.12357857078313828, 0.05596024543046951, -0.12189753353595734, 0.1458878517150879, 0.13413557410240173, -0.05317119136452675, 0.14458951354026794, -0.03716878220438957, -0.08209243416786194, -0.04329831525683403, 0.03305334970355034, 0.03183460980653763, 0.11292722821235657, -0.09258798509836197, -0.015919959172606468, 0.023859534412622452, 0.006298945285379887, 0.014593049883842468, -0.19748693704605103, -0.021644799038767815, 0.03729412704706192, -0.052588239312171936, -0.004088051617145538, -0.01409463956952095, -0.005145531613379717, 0.0978650152683258, 0.0019946310203522444, -0.048890527337789536, 0.02708890475332737, -0.009449265897274017, -0.08845166862010956, 0.21070018410682678, -0.07551591843366623, -0.16191041469573975, -0.1487201452255249, -0.03650655224919319, -0.0697425901889801, 0.022753650322556496, 0.02880474552512169, -0.05315525457262993, -0.02832791581749916, -0.10837166756391525, -0.017479587346315384, 0.01274053193628788, 0.036351799964904785, 0.03214847669005394, -0.013211037963628769, 0.05376601964235306, -0.08901341259479523, 0.0010935511672869325, -0.018517663702368736, -0.02334596775472164, 0.04963051900267601, 0.01313394121825695, 0.12095639109611511, 0.13242216408252716, -0.0034573739394545555, 0.014200377278029919, -0.022161295637488365, 0.23120394349098206, -0.06688625365495682, -0.021650375798344612, 0.10642419010400772, -0.03525792062282562, 0.06455990672111511, 0.13531368970870972, 0.04129911959171295, -0.08943761140108109, 0.015235628001391888, 0.00036843729321844876, -0.03727898374199867, -0.20887720584869385, -0.03815697506070137, -0.05536375194787979, 0.012016573920845985, 0.1085573211312294, 0.02184346504509449, 0.009550917893648148, 0.07656227797269821, -0.007250968832522631, 0.060909293591976166, -0.011942112818360329, 0.07418683916330338, 0.06579100340604782, 0.04896504059433937, 0.13609115779399872, -0.03473198786377907, -0.05032694339752197, 0.03322445601224899, 0.012381352484226227, 0.21483078598976135, -0.023811636492609978, 0.20358751714229584, 0.022891009226441383, 0.16320878267288208, 0.01254960149526596, 0.09658481925725937, 0.004405472427606583, -0.015130187384784222, -0.004714441951364279, -0.05427346006035805, -0.05910022556781769, 0.03103889897465706, -0.016948707401752472, 0.04835963621735573, -0.10968080908060074, 0.04682912304997444, 0.04203685373067856, 0.2462051510810852, 0.07873335480690002, -0.3690631687641144, -0.09834164381027222, 0.0242295041680336, -0.002525927033275366, -0.0377773717045784, -0.0015277620404958725, 0.1761069893836975, -0.0789685994386673, 0.04446892440319061, -0.08488747477531433, 0.07780781388282776, -0.07452265173196793, 0.011291295289993286, 0.03928222507238388, 0.07535445690155029, -0.012912076897919178, 0.05462873354554176, -0.24708375334739685, 0.2627999782562256, 0.017706112936139107, 0.055677030235528946, -0.06888391077518463, 0.0032666665501892567, 0.021436532959342003, 0.02714540995657444, 0.08371993154287338, -0.0029154003132134676, -0.07003571838140488, -0.1931576430797577, -0.14095188677310944, 0.01755749061703682, 0.11135894805192947, -0.059638284146785736, 0.1317857950925827, -0.02324455976486206, -0.017792630940675735, 0.04532989114522934, -0.005950724706053734, -0.05573585256934166, -0.11187880486249924, 0.027757665142416954, 0.02742382325232029, -0.010148625820875168, -0.07533854246139526, -0.11633220314979553, -0.07656475156545639, 0.15858864784240723, -0.0521361269056797, -0.04583440348505974, -0.12909723818302155, 0.0639147087931633, 0.10936560481786728, -0.09734611213207245, 0.03353460505604744, 0.005774836987257004, 0.12712745368480682, 0.005027282051742077, -0.05487631633877754, 0.09366197139024734, -0.07110445201396942, -0.19906571507453918, -0.05255942419171333, 0.14577330648899078, 0.017272869125008583, 0.053338561207056046, -0.010615913197398186, 0.031101617962121964, -0.018681198358535767, -0.0836728885769844, 0.03152112290263176, 0.005082009360194206, 0.061123304069042206, 0.030322931706905365, -0.03226467967033386, 0.0002463066775817424, -0.049051184207201004, -0.023305559530854225, 0.17834316194057465, 0.2898728847503662, -0.07994955033063889, 0.02225821278989315, 0.053377311676740646, -0.04956218600273132, -0.19093996286392212, 0.012761098332703114, 0.0722663626074791, 0.043487243354320526, 0.0178923849016428, -0.14553457498550415, 0.05396148934960365, 0.08733486384153366, -0.02119491621851921, 0.10856319218873978, -0.3024442195892334, -0.13804267346858978, 0.07599130272865295, 0.1432476043701172, 0.09388387203216553, -0.162247896194458, -0.06493847817182541, -0.02979063242673874, -0.13369961082935333, 0.12697625160217285, -0.06908344477415085, 0.11504648625850677, -0.013623976148664951, 0.04039536416530609, 0.017279306426644325, -0.06801941245794296, 0.15001320838928223, 0.011057420633733273, 0.06596098095178604, -0.042871490120887756, -0.020387640222907066, 0.08400356769561768, -0.0713367834687233, 0.028627485036849976, -0.10815343260765076, 0.045898906886577606, -0.1023423969745636, -0.02621549740433693, -0.0704566165804863, 0.010873041115701199, -0.04152897372841835, -0.03966435045003891, -0.011270827613770962, 0.033078212291002274, 0.06143873929977417, -0.0010678695980459452, 0.13981367647647858, 0.01704990491271019, 0.1328538954257965, 0.13218393921852112, 0.09759331494569778, -0.033221300691366196, -0.013058389537036419, -0.01689888723194599, -0.02743169292807579, 0.04247092455625534, -0.11298932880163193, 0.030656347051262856, 0.13164667785167694, 0.02922745980322361, 0.1412946581840515, 0.06600219011306763, -0.037744417786598206, 0.021590648218989372, 0.0682421624660492, -0.17717909812927246, -0.11355537921190262, -0.019236920401453972, -0.02986300364136696, -0.14535170793533325, 0.03825576230883598, 0.1413125842809677, -0.05788392201066017, -0.009231277741491795, -0.023668095469474792, 0.01841612346470356, -0.007546829991042614, 0.19389447569847107, 0.05681539326906204, 0.06936119496822357, -0.0954485759139061, 0.06371750682592392, 0.05177006125450134, -0.08990246057510376, 0.02204667404294014, 0.06011072173714638, -0.09844353795051575, -0.028892436996102333, 0.02871107868850231, 0.13612905144691467, -0.02954890951514244, -0.034773070365190506, -0.13813148438930511, -0.0988236740231514, 0.05935283750295639, 0.15603473782539368, 0.07056581228971481, 0.03901567682623863, -0.012528026476502419, -0.0009112288244068623, -0.11732510477304459, 0.12155482172966003, 0.06083208695054054, 0.08968799561262131, -0.1414102464914322, 0.13934685289859772, -0.025929410010576248, 0.02866298332810402, -0.019941486418247223, 0.034732166677713394, -0.11062565445899963, -0.017316697165369987, -0.10602547228336334, 0.011193067766726017, -0.05146258696913719, -0.01069188117980957, -0.015473566018044949, -0.062081966549158096, -0.05987539887428284, 0.005454647820442915, -0.09645608812570572, -0.04702111333608627, -0.0017685589846223593, 0.0342819020152092, -0.12465321272611618, -0.0345555916428566, 0.02939765527844429, -0.10578230023384094, 0.09014414995908737, 0.052039653062820435, 0.02184610813856125, 0.02760908007621765, -0.09041351079940796, 0.023912977427244186, 0.04320904612541199, 0.009191179648041725, 0.04873274639248848, -0.11135503649711609, -0.007791855838149786, -0.0032787066884338856, 0.029607834294438362, 0.02300669439136982, 0.09853484481573105, -0.10903478413820267, 0.004357147496193647, -0.027547238394618034, -0.03224201500415802, -0.06861140578985214, 0.04947389289736748, 0.07682342082262039, 0.017401007935404778, 0.1717808097600937, -0.07680965960025787, 0.03426443785429001, -0.21709349751472473, -0.0070903124287724495, 0.00013074376329313964, -0.11527346074581146, -0.09027007967233658, -0.040039774030447006, 0.0834982618689537, -0.05362183228135109, 0.11068233847618103, -0.04676975682377815, 0.04035300016403198, 0.01608884148299694, 0.00040329841431230307, 0.03211219236254692, 0.04390078783035278, 0.19964535534381866, 0.04048262909054756, -0.04258442297577858, 0.05315329134464264, 0.02993023954331875, 0.09254911541938782, 0.09042096883058548, 0.19149065017700195, 0.1184980571269989, 0.008557865396142006, 0.10269073396921158, 0.052322208881378174, -0.04785262793302536, -0.14669857919216156, 0.044459544122219086, -0.06777367740869522, 0.09856559336185455, -0.004792721942067146, 0.21348538994789124, 0.08898928016424179, -0.1808505654335022, 0.009021664969623089, -0.0349578931927681, -0.0966128334403038, -0.0981852263212204, -0.07993890345096588, -0.09889477491378784, -0.1090167984366417, -0.0031146365217864513, -0.13052068650722504, 0.007754502352327108, 0.10836312919855118, 0.021443480625748634, 0.004570317454636097, 0.1621304154396057, 0.0381263792514801, 0.02762257307767868, 0.05704662203788757, 0.015781749039888382, 0.0013618158409371972, -0.03908440098166466, -0.0984811782836914, 0.020528048276901245, -0.005445846356451511, 0.05654677748680115, -0.048976145684719086, -0.020570451393723488, 0.06303168088197708, 0.004549408797174692, -0.12222123891115189, 0.002735316986218095, 0.013202570378780365, 0.0634520947933197, 0.04705095663666725, 0.01827564835548401, 0.0044250148348510265, -0.02009090594947338, 0.19630365073680878, -0.06701144576072693, -0.04310893267393112, -0.11691483110189438, 0.15339133143424988, -0.00017559003026690334, -0.013362372294068336, 0.03696916997432709, -0.08535036444664001, 0.021375443786382675, 0.17830593883991241, 0.17342770099639893, -0.09114038199186325, -0.0012044002069160342, 0.00878489576280117, -0.0070243398658931255, -0.011649439111351967, 0.08977258205413818, 0.09170010685920715, 0.054005447775125504, -0.0804993137717247, -0.04390113428235054, -0.04088735207915306, -0.025262879207730293, -0.02675638161599636, 0.05871894210577011, 0.005344856064766645, 0.011793076060712337, -0.04494355991482735, 0.055917587131261826, -0.04616673290729523, -0.08805656433105469, 0.027696900069713593, -0.22320005297660828, -0.17678242921829224, -0.028817588463425636, 0.07706073671579361, -0.004263261798769236, 0.052336808294057846, -0.01447346806526184, 0.00040438296855427325, 0.06808830797672272, -0.0210904348641634, -0.06308627873659134, -0.05019048601388931, 0.05795660614967346, -0.0835476815700531, 0.18822041153907776, -0.03550385683774948, 0.04776233062148094, 0.12973368167877197, 0.03759852796792984, -0.10589234530925751, 0.04838767275214195, 0.07060621678829193, -0.059903841465711594, 0.04848337918519974, 0.14104779064655304, -0.04691825434565544, 0.1021137610077858, 0.05320560187101364, -0.11021129786968231, 0.004793944768607616, -0.03454664722084999, -0.036172181367874146, -0.04389107599854469, -0.04262937232851982, -0.04489537328481674, 0.15442606806755066, 0.19158229231834412, -0.05448246747255325, -0.0029260111041367054, -0.03632216155529022, 0.012391475960612297, 0.05273204669356346, 0.07800042629241943, -0.03897577524185181, -0.2708432674407959, 0.005288603715598583, 0.03932180255651474, 0.017650026828050613, -0.2836294174194336, -0.08622459322214127, -0.010955940932035446, -0.04677879437804222, -0.09472939372062683, 0.10323774069547653, 0.08181723952293396, 0.041139889508485794, -0.0540897361934185, -0.03910079970955849, -0.07118270546197891, 0.15706594288349152, -0.14952470362186432, -0.09962835162878036 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # segformer-b3-from-scratch-final This model is a fine-tuned version of [](https://huggingface.co/) on the samitizerxu/kelp_data_rgbagg_swin_nir_int_cleaned dataset. It achieves the following results on the evaluation set: - Iou Kelp: 0.0073 - Loss: 0.9864 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.001 - train_batch_size: 8 - eval_batch_size: 8 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: cosine - num_epochs: 60 ### Training results | Training Loss | Epoch | Step | Iou Kelp | Validation Loss | |:-------------:|:-----:|:-----:|:--------:|:---------------:| | 0.9993 | 0.18 | 100 | 0.0069 | 0.9867 | | 0.9945 | 0.37 | 200 | 0.0076 | 0.9855 | | 0.9991 | 0.55 | 300 | 0.0069 | 0.9867 | | 0.999 | 0.74 | 400 | 0.0066 | 0.9870 | | 0.9959 | 0.92 | 500 | 0.0071 | 0.9864 | | 0.9965 | 1.11 | 600 | 0.0066 | 0.9871 | | 0.9764 | 1.29 | 700 | 0.0066 | 0.9871 | | 0.9951 | 1.48 | 800 | 0.0066 | 0.9871 | | 0.9999 | 1.66 | 900 | 0.0066 | 0.9870 | | 0.9878 | 1.85 | 1000 | 0.0066 | 0.9871 | | 0.9978 | 2.03 | 1100 | 0.0066 | 0.9871 | | 0.975 | 2.21 | 1200 | 0.0069 | 0.9868 | | 0.9957 | 2.4 | 1300 | 0.0073 | 0.9859 | | 0.9914 | 2.58 | 1400 | 0.0079 | 0.9860 | | 0.9928 | 2.77 | 1500 | 0.0074 | 0.9859 | | 0.9994 | 2.95 | 1600 | 0.0004 | 0.9863 | | 0.995 | 3.14 | 1700 | 0.0101 | 0.9860 | | 0.9963 | 3.32 | 1800 | 0.0 | 0.9872 | | 0.9972 | 3.51 | 1900 | 0.0074 | 0.9858 | | 0.9959 | 3.69 | 2000 | 0.0076 | 0.9859 | | 0.9941 | 3.87 | 2100 | 0.0073 | 0.9859 | | 0.992 | 4.06 | 2200 | 0.0002 | 0.9951 | | 0.9903 | 4.24 | 2300 | 0.0073 | 0.9859 | | 0.9989 | 4.43 | 2400 | 0.0066 | 0.9871 | | 0.9999 | 4.61 | 2500 | 0.0073 | 0.9866 | | 0.9946 | 4.8 | 2600 | 0.0073 | 0.9859 | | 0.9959 | 4.98 | 2700 | 0.0073 | 0.9859 | | 0.9971 | 5.17 | 2800 | 0.0079 | 0.9863 | | 0.9949 | 5.35 | 2900 | 0.0074 | 0.9859 | | 0.9846 | 5.54 | 3000 | 0.0073 | 0.9859 | | 0.9941 | 5.72 | 3100 | 0.0074 | 0.9859 | | 0.9867 | 5.9 | 3200 | 0.0074 | 0.9858 | | 0.9857 | 6.09 | 3300 | 0.0074 | 0.9861 | | 0.9986 | 6.27 | 3400 | 0.0074 | 0.9859 | | 0.9927 | 6.46 | 3500 | 0.0074 | 0.9860 | | 0.998 | 6.64 | 3600 | 0.0075 | 0.9858 | | 0.9971 | 6.83 | 3700 | 0.0074 | 0.9859 | | 0.9951 | 7.01 | 3800 | 0.0074 | 0.9859 | | 0.9998 | 7.2 | 3900 | 0.0074 | 0.9861 | | 0.995 | 7.38 | 4000 | 0.0075 | 0.9858 | | 0.9912 | 7.56 | 4100 | 0.0072 | 0.9861 | | 0.9995 | 7.75 | 4200 | 0.0074 | 0.9858 | | 0.9934 | 7.93 | 4300 | 0.0074 | 0.9860 | | 0.9885 | 8.12 | 4400 | 0.0074 | 0.9860 | | 0.9937 | 8.3 | 4500 | 0.0075 | 0.9857 | | 0.9954 | 8.49 | 4600 | 0.0075 | 0.9857 | | 0.9794 | 8.67 | 4700 | 0.0074 | 0.9858 | | 0.9967 | 8.86 | 4800 | 0.0075 | 0.9857 | | 0.9954 | 9.04 | 4900 | 0.0074 | 0.9862 | | 0.9966 | 9.23 | 5000 | 0.0074 | 0.9859 | | 0.9953 | 9.41 | 5100 | 0.0074 | 0.9859 | | 0.9961 | 9.59 | 5200 | 0.0074 | 0.9859 | | 0.993 | 9.78 | 5300 | 0.0075 | 0.9858 | | 0.9993 | 9.96 | 5400 | 0.0070 | 0.9870 | | 0.995 | 10.15 | 5500 | 0.0032 | 0.9933 | | 0.9945 | 10.33 | 5600 | 0.0061 | 0.9884 | | 0.9738 | 10.52 | 5700 | 0.0069 | 0.9866 | | 0.9983 | 10.7 | 5800 | 0.0067 | 0.9869 | | 0.9975 | 10.89 | 5900 | 0.0076 | 0.9854 | | 0.9925 | 11.07 | 6000 | 0.0086 | 0.9839 | | 0.9821 | 11.25 | 6100 | 0.0092 | 0.9822 | | 0.9972 | 11.44 | 6200 | 0.0107 | 0.9787 | | 0.9802 | 11.62 | 6300 | 0.0109 | 0.9781 | | 1.0 | 11.81 | 6400 | 0.0076 | 0.9854 | | 0.9922 | 11.99 | 6500 | 0.0108 | 0.9793 | | 0.9915 | 12.18 | 6600 | 0.0108 | 0.9799 | | 0.9963 | 12.36 | 6700 | 0.0075 | 0.9857 | | 0.9966 | 12.55 | 6800 | 0.0075 | 0.9859 | | 0.9978 | 12.73 | 6900 | 0.0069 | 0.9870 | | 0.9847 | 12.92 | 7000 | 0.0074 | 0.9860 | | 0.9972 | 13.1 | 7100 | 0.0072 | 0.9862 | | 0.9868 | 13.28 | 7200 | 0.0071 | 0.9865 | | 0.9961 | 13.47 | 7300 | 0.0072 | 0.9864 | | 0.9845 | 13.65 | 7400 | 0.0071 | 0.9865 | | 0.9974 | 13.84 | 7500 | 0.0074 | 0.9862 | | 0.9906 | 14.02 | 7600 | 0.0076 | 0.9847 | | 0.9999 | 14.21 | 7700 | 0.0075 | 0.9860 | | 0.9821 | 14.39 | 7800 | 0.0074 | 0.9860 | | 0.9976 | 14.58 | 7900 | 0.0105 | 0.9795 | | 0.9871 | 14.76 | 8000 | 0.0103 | 0.9803 | | 0.991 | 14.94 | 8100 | 0.0102 | 0.9805 | | 0.9903 | 15.13 | 8200 | 0.0104 | 0.9799 | | 0.995 | 15.31 | 8300 | 0.0074 | 0.9861 | | 0.9981 | 15.5 | 8400 | 0.0073 | 0.9863 | | 0.9985 | 15.68 | 8500 | 0.0073 | 0.9863 | | 0.9973 | 15.87 | 8600 | 0.0074 | 0.9862 | | 0.989 | 16.05 | 8700 | 0.0073 | 0.9863 | | 0.9938 | 16.24 | 8800 | 0.0074 | 0.9860 | | 0.9951 | 16.42 | 8900 | 0.0106 | 0.9786 | | 0.9921 | 16.61 | 9000 | 0.0092 | 0.9824 | | 0.9971 | 16.79 | 9100 | 0.0083 | 0.9846 | | 0.9846 | 16.97 | 9200 | 0.0087 | 0.9838 | | 0.9849 | 17.16 | 9300 | 0.0095 | 0.9820 | | 0.9851 | 17.34 | 9400 | 0.0096 | 0.9818 | | 0.9902 | 17.53 | 9500 | 0.0099 | 0.9811 | | 0.9889 | 17.71 | 9600 | 0.0075 | 0.9860 | | 0.9782 | 17.9 | 9700 | 0.0075 | 0.9908 | | 0.999 | 18.08 | 9800 | 0.0074 | 0.9862 | | 0.9878 | 18.27 | 9900 | 0.0073 | 0.9862 | | 0.999 | 18.45 | 10000 | 0.0074 | 0.9862 | | 1.0 | 18.63 | 10100 | 0.0074 | 0.9861 | | 0.9951 | 18.82 | 10200 | 0.0075 | 0.9859 | | 0.9892 | 19.0 | 10300 | 0.0073 | 0.9861 | | 0.9853 | 19.19 | 10400 | 0.0074 | 0.9859 | | 0.9959 | 19.37 | 10500 | 0.0074 | 0.9859 | | 0.9999 | 19.56 | 10600 | 0.0073 | 0.9861 | | 0.9872 | 19.74 | 10700 | 0.0074 | 0.9859 | | 0.9939 | 19.93 | 10800 | 0.0074 | 0.9861 | | 0.9924 | 20.11 | 10900 | 0.0073 | 0.9862 | | 0.9993 | 20.3 | 11000 | 0.0074 | 0.9860 | | 0.9934 | 20.48 | 11100 | 0.0075 | 0.9858 | | 0.9976 | 20.66 | 11200 | 0.0074 | 0.9859 | | 0.9878 | 20.85 | 11300 | 0.0074 | 0.9859 | | 0.9955 | 21.03 | 11400 | 0.0074 | 0.9859 | | 0.9878 | 21.22 | 11500 | 0.0075 | 0.9859 | | 0.999 | 21.4 | 11600 | 0.0074 | 0.9859 | | 0.9945 | 21.59 | 11700 | 0.0074 | 0.9861 | | 0.994 | 21.77 | 11800 | 0.0075 | 0.9859 | | 0.9848 | 21.96 | 11900 | 0.0075 | 0.9859 | | 0.9998 | 22.14 | 12000 | 0.0075 | 0.9859 | | 0.9826 | 22.32 | 12100 | 0.0075 | 0.9859 | | 0.999 | 22.51 | 12200 | 0.0074 | 0.9861 | | 0.9941 | 22.69 | 12300 | 0.0073 | 0.9863 | | 0.9933 | 22.88 | 12400 | 0.0074 | 0.9862 | | 0.9935 | 23.06 | 12500 | 0.0074 | 0.9862 | | 0.9992 | 23.25 | 12600 | 0.0073 | 0.9863 | | 0.9943 | 23.43 | 12700 | 0.0073 | 0.9863 | | 0.9777 | 23.62 | 12800 | 0.0075 | 0.9858 | | 0.9977 | 23.8 | 12900 | 0.0073 | 0.9862 | | 0.9925 | 23.99 | 13000 | 0.0074 | 0.9861 | | 0.9866 | 24.17 | 13100 | 0.0073 | 0.9863 | | 0.9979 | 24.35 | 13200 | 0.0073 | 0.9862 | | 0.9819 | 24.54 | 13300 | 0.0073 | 0.9864 | | 0.966 | 24.72 | 13400 | 0.0073 | 0.9864 | | 0.998 | 24.91 | 13500 | 0.0073 | 0.9863 | | 0.9969 | 25.09 | 13600 | 0.0073 | 0.9863 | | 0.9881 | 25.28 | 13700 | 0.0073 | 0.9863 | | 0.9701 | 25.46 | 13800 | 0.0073 | 0.9864 | | 0.9963 | 25.65 | 13900 | 0.0073 | 0.9863 | | 0.9885 | 25.83 | 14000 | 0.0073 | 0.9863 | | 0.9904 | 26.01 | 14100 | 0.0073 | 0.9864 | | 0.9976 | 26.2 | 14200 | 0.0074 | 0.9862 | | 0.995 | 26.38 | 14300 | 0.0073 | 0.9863 | | 0.9886 | 26.57 | 14400 | 0.0073 | 0.9864 | | 0.9735 | 26.75 | 14500 | 0.0073 | 0.9863 | | 0.988 | 26.94 | 14600 | 0.0073 | 0.9864 | | 0.9854 | 27.12 | 14700 | 0.0073 | 0.9864 | | 0.9947 | 27.31 | 14800 | 0.0073 | 0.9864 | | 0.9944 | 27.49 | 14900 | 0.0073 | 0.9864 | | 0.9935 | 27.68 | 15000 | 0.0073 | 0.9862 | | 0.9887 | 27.86 | 15100 | 0.0073 | 0.9863 | | 0.9958 | 28.04 | 15200 | 0.0073 | 0.9862 | | 0.9994 | 28.23 | 15300 | 0.0073 | 0.9863 | | 0.9953 | 28.41 | 15400 | 0.0073 | 0.9868 | | 0.9798 | 28.6 | 15500 | 0.0073 | 0.9863 | | 0.9867 | 28.78 | 15600 | 0.0073 | 0.9863 | | 0.9903 | 28.97 | 15700 | 0.0073 | 0.9863 | | 0.9943 | 29.15 | 15800 | 0.0073 | 0.9864 | | 0.9725 | 29.34 | 15900 | 0.0072 | 0.9864 | | 0.9987 | 29.52 | 16000 | 0.0073 | 0.9864 | | 0.9871 | 29.7 | 16100 | 0.0072 | 0.9864 | | 0.992 | 29.89 | 16200 | 0.0072 | 0.9864 | | 0.996 | 30.07 | 16300 | 0.0073 | 0.9864 | | 0.9998 | 30.26 | 16400 | 0.0073 | 0.9864 | | 0.9964 | 30.44 | 16500 | 0.0074 | 0.9859 | | 0.9992 | 30.63 | 16600 | 0.0075 | 0.9858 | | 0.9946 | 30.81 | 16700 | 0.0074 | 0.9861 | | 0.9911 | 31.0 | 16800 | 0.0075 | 0.9859 | | 0.9878 | 31.18 | 16900 | 0.0075 | 0.9859 | | 0.9826 | 31.37 | 17000 | 0.0075 | 0.9859 | | 0.9894 | 31.55 | 17100 | 0.0075 | 0.9859 | | 0.9887 | 31.73 | 17200 | 0.0075 | 0.9860 | | 0.9962 | 31.92 | 17300 | 0.0073 | 0.9862 | | 0.9937 | 32.1 | 17400 | 0.0073 | 0.9863 | | 0.9828 | 32.29 | 17500 | 0.0073 | 0.9863 | | 0.993 | 32.47 | 17600 | 0.0073 | 0.9864 | | 0.9975 | 32.66 | 17700 | 0.0073 | 0.9864 | | 0.994 | 32.84 | 17800 | 0.0073 | 0.9864 | | 0.9894 | 33.03 | 17900 | 0.0073 | 0.9862 | | 0.9938 | 33.21 | 18000 | 0.0073 | 0.9863 | | 0.9711 | 33.39 | 18100 | 0.0073 | 0.9863 | | 0.9896 | 33.58 | 18200 | 0.0073 | 0.9864 | | 0.9907 | 33.76 | 18300 | 0.0073 | 0.9864 | | 0.9934 | 33.95 | 18400 | 0.0073 | 0.9864 | | 0.9723 | 34.13 | 18500 | 0.0073 | 0.9864 | | 0.9842 | 34.32 | 18600 | 0.0073 | 0.9864 | | 0.9955 | 34.5 | 18700 | 0.0073 | 0.9864 | | 0.9824 | 34.69 | 18800 | 0.0073 | 0.9864 | | 0.9949 | 34.87 | 18900 | 0.0073 | 0.9864 | | 0.9943 | 35.06 | 19000 | 0.0073 | 0.9864 | | 0.9992 | 35.24 | 19100 | 0.0073 | 0.9864 | | 0.9843 | 35.42 | 19200 | 0.0073 | 0.9864 | | 0.9785 | 35.61 | 19300 | 0.0073 | 0.9864 | | 0.9999 | 35.79 | 19400 | 0.0073 | 0.9864 | | 0.9727 | 35.98 | 19500 | 0.0073 | 0.9864 | | 0.9949 | 36.16 | 19600 | 0.0073 | 0.9864 | | 0.9949 | 36.35 | 19700 | 0.0073 | 0.9864 | | 0.9887 | 36.53 | 19800 | 0.0073 | 0.9864 | | 0.9736 | 36.72 | 19900 | 0.0073 | 0.9864 | | 0.9966 | 36.9 | 20000 | 0.0073 | 0.9864 | | 0.9984 | 37.08 | 20100 | 0.0073 | 0.9864 | | 0.993 | 37.27 | 20200 | 0.0073 | 0.9864 | | 0.9998 | 37.45 | 20300 | 0.0073 | 0.9864 | | 0.9972 | 37.64 | 20400 | 0.0073 | 0.9864 | | 0.986 | 37.82 | 20500 | 0.0073 | 0.9864 | | 0.9914 | 38.01 | 20600 | 0.0073 | 0.9864 | | 0.9954 | 38.19 | 20700 | 0.0073 | 0.9864 | | 0.9764 | 38.38 | 20800 | 0.0073 | 0.9864 | | 0.9953 | 38.56 | 20900 | 0.0073 | 0.9864 | | 0.9837 | 38.75 | 21000 | 0.0073 | 0.9864 | | 0.9665 | 38.93 | 21100 | 0.0073 | 0.9864 | | 0.9964 | 39.11 | 21200 | 0.0073 | 0.9864 | | 0.9935 | 39.3 | 21300 | 0.0073 | 0.9864 | | 0.9466 | 39.48 | 21400 | 0.0073 | 0.9864 | | 0.9853 | 39.67 | 21500 | 0.0073 | 0.9864 | | 0.9678 | 39.85 | 21600 | 0.0073 | 0.9864 | | 0.995 | 40.04 | 21700 | 0.0073 | 0.9864 | | 0.9987 | 40.22 | 21800 | 0.0073 | 0.9864 | | 0.9935 | 40.41 | 21900 | 0.0073 | 0.9864 | | 0.991 | 40.59 | 22000 | 0.0073 | 0.9864 | | 0.999 | 40.77 | 22100 | 0.0073 | 0.9864 | | 0.9985 | 40.96 | 22200 | 0.0073 | 0.9864 | | 0.9954 | 41.14 | 22300 | 0.0073 | 0.9864 | | 0.9894 | 41.33 | 22400 | 0.0073 | 0.9864 | | 0.9851 | 41.51 | 22500 | 0.0073 | 0.9864 | | 0.9882 | 41.7 | 22600 | 0.0073 | 0.9864 | | 0.9999 | 41.88 | 22700 | 0.0073 | 0.9864 | | 0.9901 | 42.07 | 22800 | 0.0073 | 0.9864 | | 0.9853 | 42.25 | 22900 | 0.0073 | 0.9864 | | 0.9868 | 42.44 | 23000 | 0.0073 | 0.9864 | | 0.9973 | 42.62 | 23100 | 0.0073 | 0.9864 | | 0.9979 | 42.8 | 23200 | 0.0073 | 0.9865 | | 0.9867 | 42.99 | 23300 | 0.0073 | 0.9864 | | 0.9994 | 43.17 | 23400 | 0.0073 | 0.9864 | | 0.9984 | 43.36 | 23500 | 0.0073 | 0.9865 | | 0.9974 | 43.54 | 23600 | 0.0073 | 0.9865 | | 0.9999 | 43.73 | 23700 | 0.0073 | 0.9864 | | 0.9669 | 43.91 | 23800 | 0.0073 | 0.9864 | | 0.9925 | 44.1 | 23900 | 0.0073 | 0.9864 | | 0.9961 | 44.28 | 24000 | 0.0073 | 0.9864 | | 0.9815 | 44.46 | 24100 | 0.0073 | 0.9864 | | 0.9968 | 44.65 | 24200 | 0.0073 | 0.9864 | | 0.9964 | 44.83 | 24300 | 0.0073 | 0.9864 | | 0.9929 | 45.02 | 24400 | 0.0073 | 0.9864 | | 0.9712 | 45.2 | 24500 | 0.0073 | 0.9864 | | 0.9884 | 45.39 | 24600 | 0.0073 | 0.9864 | | 0.9897 | 45.57 | 24700 | 0.0073 | 0.9864 | | 0.9862 | 45.76 | 24800 | 0.0073 | 0.9865 | | 0.9768 | 45.94 | 24900 | 0.0073 | 0.9865 | | 0.9965 | 46.13 | 25000 | 0.0073 | 0.9865 | | 0.9996 | 46.31 | 25100 | 0.0073 | 0.9864 | | 0.9887 | 46.49 | 25200 | 0.0073 | 0.9864 | | 0.9991 | 46.68 | 25300 | 0.0073 | 0.9864 | | 0.984 | 46.86 | 25400 | 0.0073 | 0.9864 | | 0.983 | 47.05 | 25500 | 0.0073 | 0.9864 | | 0.9997 | 47.23 | 25600 | 0.0073 | 0.9864 | | 0.9923 | 47.42 | 25700 | 0.0073 | 0.9865 | | 0.9962 | 47.6 | 25800 | 0.0073 | 0.9864 | | 0.9747 | 47.79 | 25900 | 0.0073 | 0.9864 | | 0.9981 | 47.97 | 26000 | 0.0073 | 0.9864 | | 0.9936 | 48.15 | 26100 | 0.0073 | 0.9864 | | 0.9976 | 48.34 | 26200 | 0.0073 | 0.9864 | | 0.9601 | 48.52 | 26300 | 0.0073 | 0.9865 | | 0.9881 | 48.71 | 26400 | 0.0073 | 0.9864 | | 0.9919 | 48.89 | 26500 | 0.0073 | 0.9864 | | 0.9748 | 49.08 | 26600 | 0.0073 | 0.9864 | | 0.9862 | 49.26 | 26700 | 0.0073 | 0.9864 | | 0.9935 | 49.45 | 26800 | 0.0073 | 0.9864 | | 0.9402 | 49.63 | 26900 | 0.0073 | 0.9864 | | 0.9982 | 49.82 | 27000 | 0.0073 | 0.9864 | | 0.9619 | 50.0 | 27100 | 0.0073 | 0.9864 | | 0.9935 | 50.18 | 27200 | 0.0073 | 0.9864 | | 0.9962 | 50.37 | 27300 | 0.0073 | 0.9864 | | 0.9888 | 50.55 | 27400 | 0.0073 | 0.9864 | | 0.9956 | 50.74 | 27500 | 0.0073 | 0.9864 | | 0.9981 | 50.92 | 27600 | 0.0073 | 0.9864 | | 0.9992 | 51.11 | 27700 | 0.0073 | 0.9865 | | 0.9613 | 51.29 | 27800 | 0.0073 | 0.9864 | | 0.9721 | 51.48 | 27900 | 0.0072 | 0.9865 | | 0.9938 | 51.66 | 28000 | 0.0073 | 0.9865 | | 0.9998 | 51.85 | 28100 | 0.0073 | 0.9864 | | 0.9981 | 52.03 | 28200 | 0.0073 | 0.9864 | | 0.9793 | 52.21 | 28300 | 0.0073 | 0.9864 | | 0.9962 | 52.4 | 28400 | 0.0073 | 0.9864 | | 0.9728 | 52.58 | 28500 | 0.0073 | 0.9864 | | 0.9965 | 52.77 | 28600 | 0.0073 | 0.9864 | | 0.9937 | 52.95 | 28700 | 0.0073 | 0.9864 | | 0.9942 | 53.14 | 28800 | 0.0073 | 0.9864 | | 0.9902 | 53.32 | 28900 | 0.0073 | 0.9864 | | 0.9992 | 53.51 | 29000 | 0.0073 | 0.9864 | | 0.9954 | 53.69 | 29100 | 0.0073 | 0.9864 | | 0.991 | 53.87 | 29200 | 0.0073 | 0.9864 | | 0.9955 | 54.06 | 29300 | 0.0073 | 0.9864 | | 0.9978 | 54.24 | 29400 | 0.0073 | 0.9864 | | 0.9998 | 54.43 | 29500 | 0.0073 | 0.9864 | | 0.9716 | 54.61 | 29600 | 0.0073 | 0.9864 | | 0.9891 | 54.8 | 29700 | 0.0073 | 0.9864 | | 0.9984 | 54.98 | 29800 | 0.0073 | 0.9864 | | 0.9756 | 55.17 | 29900 | 0.0073 | 0.9864 | | 0.9901 | 55.35 | 30000 | 0.0073 | 0.9864 | | 0.9866 | 55.54 | 30100 | 0.0073 | 0.9864 | | 0.9908 | 55.72 | 30200 | 0.0073 | 0.9864 | | 0.977 | 55.9 | 30300 | 0.0073 | 0.9864 | | 0.9882 | 56.09 | 30400 | 0.0073 | 0.9864 | | 0.9903 | 56.27 | 30500 | 0.0073 | 0.9864 | | 0.9819 | 56.46 | 30600 | 0.0073 | 0.9864 | | 0.9883 | 56.64 | 30700 | 0.0073 | 0.9864 | | 0.9922 | 56.83 | 30800 | 0.0073 | 0.9864 | | 0.9788 | 57.01 | 30900 | 0.0073 | 0.9864 | | 0.9756 | 57.2 | 31000 | 0.0073 | 0.9864 | | 0.9955 | 57.38 | 31100 | 0.0073 | 0.9864 | | 0.9925 | 57.56 | 31200 | 0.0073 | 0.9864 | | 0.9976 | 57.75 | 31300 | 0.0073 | 0.9864 | | 0.9938 | 57.93 | 31400 | 0.0073 | 0.9864 | | 0.9905 | 58.12 | 31500 | 0.0073 | 0.9864 | | 0.9819 | 58.3 | 31600 | 0.0073 | 0.9864 | | 0.9827 | 58.49 | 31700 | 0.0073 | 0.9864 | | 0.9927 | 58.67 | 31800 | 0.0073 | 0.9864 | | 0.9953 | 58.86 | 31900 | 0.0073 | 0.9864 | | 0.9937 | 59.04 | 32000 | 0.0073 | 0.9864 | | 0.9961 | 59.23 | 32100 | 0.0073 | 0.9864 | | 0.9886 | 59.41 | 32200 | 0.0073 | 0.9864 | | 0.9906 | 59.59 | 32300 | 0.0073 | 0.9864 | | 0.9811 | 59.78 | 32400 | 0.0073 | 0.9864 | | 0.9977 | 59.96 | 32500 | 0.0073 | 0.9864 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.2.0 - Datasets 2.17.0 - Tokenizers 0.15.1
{"tags": ["vision", "image-segmentation", "generated_from_trainer"], "model-index": [{"name": "segformer-b3-from-scratch-final", "results": []}]}
image-segmentation
samitizerxu/segformer-b3-from-scratch-final
[ "transformers", "safetensors", "segformer", "vision", "image-segmentation", "generated_from_trainer", "endpoints_compatible", "region:us" ]
2024-02-10T15:12:23+00:00
[]
[]
TAGS #transformers #safetensors #segformer #vision #image-segmentation #generated_from_trainer #endpoints_compatible #region-us
segformer-b3-from-scratch-final =============================== This model is a fine-tuned version of [](URL on the samitizerxu/kelp\_data\_rgbagg\_swin\_nir\_int\_cleaned dataset. It achieves the following results on the evaluation set: * Iou Kelp: 0.0073 * Loss: 0.9864 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 0.001 * train\_batch\_size: 8 * eval\_batch\_size: 8 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: cosine * num\_epochs: 60 ### Training results ### Framework versions * Transformers 4.37.2 * Pytorch 2.2.0 * Datasets 2.17.0 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* num\\_epochs: 60", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.2.0\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #safetensors #segformer #vision #image-segmentation #generated_from_trainer #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* num\\_epochs: 60", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.2.0\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ 40, 98, 4, 30 ]
[ "passage: TAGS\n#transformers #safetensors #segformer #vision #image-segmentation #generated_from_trainer #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 8\n* eval\\_batch\\_size: 8\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: cosine\n* num\\_epochs: 60### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.2.0\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ -0.10411479324102402, -0.024431081488728523, -0.0007372432737611234, 0.0908518135547638, 0.19007684290409088, 0.02338247001171112, 0.11886221915483475, 0.07503367215394974, -0.14929988980293274, 0.036451105028390884, 0.11009279638528824, 0.12325144559144974, 0.0005903051933273673, 0.12402134388685226, -0.052381131798028946, -0.2636178433895111, 0.003956907894462347, 0.026624569669365883, -0.1192958652973175, 0.10110297054052353, 0.08882938325405121, -0.17813533544540405, 0.09114334732294083, -0.018495645374059677, -0.2787429988384247, 0.029100818559527397, 0.025015346705913544, -0.04960505664348602, 0.14140582084655762, 0.017427340149879456, 0.16098757088184357, -0.0011386952828615904, 0.1251050978899002, -0.1887296736240387, 0.01448590587824583, 0.06402226537466049, 0.020970122888684273, 0.04765007272362709, 0.042806755751371384, 0.002979532815515995, 0.07863244414329529, -0.11767889559268951, 0.08554273098707199, 0.007161071058362722, -0.14539551734924316, -0.21384993195533752, -0.06794679909944534, -0.005342388059943914, 0.1008441299200058, 0.09391631931066513, 0.00267092720605433, 0.12744803726673126, -0.08660151809453964, 0.10639353841543198, 0.26879093050956726, -0.19612064957618713, -0.09507046639919281, 0.035409051924943924, -0.002292563207447529, 0.08370207995176315, -0.1363281011581421, 0.013517259620130062, 0.08509475737810135, 0.0350724458694458, 0.10175902396440506, -0.028325237333774567, -0.11200372129678726, 0.008906246162950993, -0.13949032127857208, -0.02696003019809723, 0.09873814135789871, 0.04647519811987877, -0.027790378779172897, -0.006337453611195087, -0.07261738181114197, -0.16227521002292633, -0.07607994228601456, -0.02054227702319622, 0.07011707127094269, -0.06945349276065826, -0.1177317276597023, 0.006664707791060209, -0.12860733270645142, -0.11029320955276489, -0.03661716356873512, 0.18212930858135223, 0.04039755463600159, 0.03287303075194359, -0.058006249368190765, 0.10241173207759857, -0.059912100434303284, -0.13561317324638367, 0.013626505620777607, 0.02560056559741497, -0.04057297110557556, -0.052747633308172226, -0.07932976633310318, -0.07841064035892487, -0.003166313050314784, 0.1370917707681656, -0.10707618296146393, 0.06433214247226715, 0.043095823377370834, 0.04893208667635918, -0.127317413687706, 0.19939179718494415, -0.021135695278644562, 0.010653646662831306, -0.015102758072316647, 0.03966831788420677, -0.0012331980979070067, 0.0032565761357545853, -0.08143886178731918, -0.0025707914028316736, 0.09750522673130035, -0.021037792786955833, -0.11841075867414474, 0.05852614343166351, -0.028063738718628883, 0.021501990035176277, -0.018269360065460205, -0.08833584934473038, 0.04337797313928604, -0.016061319038271904, -0.062435440719127655, -0.029909681528806686, 0.054343629628419876, 0.02685079164803028, 0.03497881069779396, 0.13588303327560425, -0.07798677682876587, 0.05578577518463135, -0.12636150419712067, -0.07610397040843964, -0.012757443822920322, -0.03206605836749077, 0.04326928034424782, -0.110376738011837, -0.16056205332279205, -0.016030490398406982, 0.05372786894440651, -0.011163272894918919, 0.017853103578090668, -0.0523676835000515, -0.10518241673707962, 0.009152129292488098, -0.014841646887362003, 0.13747242093086243, -0.053565073758363724, 0.10559544712305069, 0.08264175802469254, 0.08106021583080292, -0.07639411091804504, 0.020269468426704407, -0.08642011880874634, 0.018593454733490944, -0.271403968334198, 0.03968346491456032, -0.05218322202563286, 0.09100261330604553, -0.03468950092792511, -0.10959239304065704, -0.011362307704985142, 0.007157036568969488, 0.0867442786693573, 0.11355079710483551, -0.20211200416088104, -0.07583627104759216, 0.1558450609445572, -0.12428684532642365, -0.12315227091312408, 0.09412126243114471, -0.05536601319909096, 0.02382722869515419, 0.08379723876714706, 0.16854536533355713, 0.045379169285297394, -0.13550972938537598, 0.0030802402179688215, -0.03549445420503616, 0.04389311745762825, -0.033756036311388016, 0.040986496955156326, 0.03878482058644295, 0.04225712642073631, 0.021469801664352417, -0.0594119131565094, 0.06605713069438934, -0.13603158295154572, -0.08912258595228195, -0.05423673987388611, -0.10681705176830292, 0.02427077107131481, 0.08599060773849487, 0.09319279342889786, -0.12740656733512878, -0.046667393296957016, 0.11863875389099121, 0.06769063323736191, -0.07248072326183319, 0.03403956815600395, -0.055443890392780304, 0.022219641134142876, -0.048957545310258865, -0.035746194422245026, -0.16517247259616852, -0.06311526149511337, -0.01308131217956543, 0.05093170702457428, 0.015289544127881527, -0.012501205317676067, 0.09977596998214722, 0.0915808379650116, -0.07749202847480774, -0.04007131606340408, -0.07717768847942352, 0.008589418604969978, -0.1255689561367035, -0.20211155712604523, -0.007252553477883339, -0.00624136533588171, 0.09324458241462708, -0.20774753391742706, 0.02982320263981819, 0.011119814589619637, 0.12591314315795898, 0.03302810341119766, -0.023589391261339188, -0.08195344358682632, 0.07731416076421738, -0.0002602175227366388, -0.06624286621809006, 0.034053124487400055, -0.014258134178817272, -0.048289116472005844, -0.06475188583135605, -0.11282213032245636, 0.1804802119731903, 0.14653490483760834, -0.19013091921806335, -0.11494096368551254, 0.009190835058689117, -0.06192038580775261, -0.026511386036872864, -0.06433066725730896, 0.011944124475121498, 0.1577848494052887, -0.029472479596734047, 0.13862521946430206, -0.06805868446826935, -0.035192493349313736, 0.024460233747959137, -0.05462170019745827, 0.044446107000112534, 0.0833190381526947, 0.0928751602768898, -0.07279413938522339, 0.1265970915555954, 0.14155009388923645, -0.1171112060546875, 0.1219625174999237, -0.05661952868103981, -0.07827936112880707, 0.005440625827759504, -0.015985414385795593, -0.0037070252001285553, 0.17202620208263397, -0.10846889019012451, -0.01803996041417122, -0.020503679290413857, 0.02748030610382557, 0.03290697559714317, -0.25621724128723145, -0.04465514048933983, 0.031404875218868256, -0.015638191252946854, 0.03761062026023865, -0.03702064976096153, 0.009844600223004818, 0.10275653004646301, -0.021785132586956024, -0.06274813413619995, 0.014966082759201527, -0.007878448814153671, -0.0647842213511467, 0.22064265608787537, -0.08076993376016617, -0.09407132118940353, -0.10268060117959976, -0.011416939087212086, -0.015496610663831234, 0.030267879366874695, 0.05842744931578636, -0.1357191801071167, -0.030727555975317955, -0.042261917144060135, 0.025073910132050514, 0.03610474616289139, 0.07112737745046616, 0.05484737828373909, 0.007239271886646748, 0.07829255610704422, -0.09214487671852112, -0.015469519421458244, -0.06610691547393799, -0.06706665456295013, 0.05464501678943634, 0.04101726785302162, 0.1391335129737854, 0.14080823957920074, -0.050457049161195755, 0.015783146023750305, -0.027298344299197197, 0.2621408700942993, -0.11070515960454941, -0.06642709672451019, 0.11997528374195099, -0.018134519457817078, 0.0317746140062809, 0.09216471761465073, 0.06513957679271698, -0.13306918740272522, 0.0150908837094903, 0.04709925502538681, -0.048256490379571915, -0.15315857529640198, -0.019657112658023834, -0.031503066420555115, -0.05419747531414032, 0.061115264892578125, 0.018742704764008522, 0.007505444344133139, 0.06796052306890488, 0.03527804836630821, 0.07083477824926376, -0.004544767085462809, 0.06169150397181511, 0.07643579691648483, 0.033811602741479874, 0.10345254838466644, -0.05055318400263786, -0.08411204814910889, 0.01594872772693634, -0.03119244985282421, 0.24389956891536713, 0.007519034668803215, 0.035870060324668884, 0.061944495886564255, 0.12204930186271667, -0.0070684743113815784, 0.06173016503453255, 0.007172773126512766, -0.08929751068353653, 0.0054962774738669395, -0.04501787945628166, -0.034072864800691605, 0.03952592983841896, -0.09873036295175552, 0.06346829235553741, -0.14072374999523163, 0.03623472899198532, 0.09057219326496124, 0.22876547276973724, 0.024074044078588486, -0.34729671478271484, -0.06437133997678757, 0.00964199099689722, -0.022205278277397156, -0.01824611984193325, 0.014509247615933418, 0.16244660317897797, -0.06528677046298981, 0.06085406616330147, -0.059783097356557846, 0.07832226902246475, -0.00540169095620513, 0.04336986690759659, 0.04483760520815849, 0.06540057808160782, 0.0038278952706605196, 0.035101305693387985, -0.27909961342811584, 0.28878095746040344, -0.0021614579018205404, 0.11271165311336517, -0.03209516406059265, -0.022854259237647057, 0.01660533808171749, 0.11402134597301483, 0.06673591583967209, -0.020990608260035515, -0.08248337358236313, -0.22738048434257507, 0.0035974893253296614, 0.0473453626036644, 0.1632213145494461, -0.0007974463514983654, 0.10263501107692719, -0.015512341633439064, 0.0049298228695988655, 0.08998926728963852, -0.044509537518024445, -0.07904613763093948, -0.06404530256986618, -0.05682959407567978, -0.008403168059885502, -0.01394875068217516, -0.08807217329740524, -0.08851027488708496, -0.11467470973730087, 0.12450607120990753, 0.009623119607567787, 0.000912059040274471, -0.12722241878509521, 0.11322426050901413, 0.06578618288040161, -0.05561976134777069, 0.06570057570934296, 0.0530911386013031, 0.08906854689121246, 0.03142933547496796, -0.047523338347673416, 0.09993588179349899, -0.07187511026859283, -0.1673957109451294, -0.05524689704179764, 0.07961572706699371, 0.045124854892492294, 0.032965198159217834, -0.00009385282464791089, 0.023357611149549484, 0.00693281227722764, -0.08992544561624527, 0.04402308538556099, 0.003788696601986885, 0.0741044282913208, 0.042341362684965134, -0.013500097207725048, 0.016977500170469284, -0.04248457029461861, -0.011518236249685287, 0.13298088312149048, 0.2970251441001892, -0.07662764191627502, -0.04777602478861809, 0.016423501074314117, -0.06011892110109329, -0.19944889843463898, 0.11902900040149689, 0.08728659898042679, -0.005809004884213209, 0.037043627351522446, -0.10784254223108292, 0.13786891102790833, 0.11227342486381531, -0.009786641225218773, 0.06778103113174438, -0.31459736824035645, -0.14580929279327393, 0.08924224227666855, 0.1916767656803131, 0.12467535585165024, -0.16052816808223724, 0.0018662912771105766, -0.057584915310144424, -0.12614886462688446, 0.07160592824220657, -0.09603433310985565, 0.10945166647434235, -0.007452905643731356, 0.07357771694660187, 0.010321281850337982, -0.06206919625401497, 0.13761921226978302, -0.0041322484612464905, 0.15892840921878815, -0.05822064355015755, -0.01676112599670887, 0.06828900426626205, -0.0473262295126915, -0.015883365646004677, 0.013578382320702076, 0.04049612954258919, -0.015038919635117054, -0.013548632152378559, -0.08375664055347443, 0.029033735394477844, -0.01979287900030613, -0.055915262550115585, -0.05588795244693756, 0.01057550311088562, 0.022650504484772682, -0.01477757841348648, 0.15471580624580383, 0.008634485304355621, 0.13344596326351166, 0.03103923238813877, 0.0571276992559433, -0.08214676380157471, -0.0639626532793045, 0.00756757752969861, -0.008259708993136883, 0.06926564127206802, -0.1457701027393341, 0.03707056865096092, 0.1354057341814041, 0.036779820919036865, 0.12364710122346878, 0.09659390151500702, -0.01742388680577278, 0.033291444182395935, 0.09206290543079376, -0.15641318261623383, -0.0970861092209816, -0.002128644147887826, -0.026804832741618156, -0.06393608450889587, 0.09287615120410919, 0.0830034464597702, -0.0929497554898262, 0.018756087869405746, -0.03686382621526718, -0.011971354484558105, -0.0846680998802185, 0.23265857994556427, 0.0701485425233841, 0.041559167206287384, -0.09636576473712921, 0.08911435306072235, 0.0014481315156444907, -0.0803002268075943, -0.004643605090677738, 0.07281329482793808, -0.06248101592063904, -0.0032647999469190836, 0.13469652831554413, 0.2116558849811554, -0.09235835075378418, -0.04606151208281517, -0.1629945933818817, -0.10713934898376465, 0.04977206513285637, 0.23302073776721954, 0.11001353710889816, 0.0008120711427181959, 0.0003899977309629321, 0.06427916139364243, -0.1409371942281723, 0.0643443614244461, 0.013931798748672009, 0.10539191961288452, -0.1685466170310974, 0.1869935244321823, 0.00745297921821475, 0.036721084266901016, -0.041225749999284744, 0.05217834562063217, -0.1280018538236618, 0.044547587633132935, -0.10272230952978134, -0.02515396848320961, 0.014350006356835365, -0.008935902267694473, 0.023284610360860825, -0.07055562734603882, -0.07151934504508972, 0.024054717272520065, -0.12164441496133804, -0.011050221510231495, 0.04303388670086861, 0.017469096928834915, -0.1237366572022438, -0.04419714957475662, 0.03784431144595146, -0.06831380724906921, 0.020641116425395012, 0.05662628635764122, 0.027448950335383415, 0.0897536352276802, -0.21260322630405426, -0.04047979786992073, 0.11663340777158737, -0.018770301714539528, 0.05945643037557602, -0.018986962735652924, -0.028872918337583542, -0.0074148778803646564, 0.1082412600517273, 0.013562631793320179, 0.056010033935308456, -0.1219199150800705, -0.007724212482571602, -0.04202227666974068, -0.08735376596450806, -0.06092905253171921, 0.009178072214126587, 0.07479339838027954, 0.019301604479551315, 0.14985208213329315, -0.09720361977815628, 0.029052134603261948, -0.20659112930297852, -0.009150318801403046, -0.001208141795359552, -0.11227023601531982, -0.08816131949424744, -0.04919268563389778, 0.08301897346973419, -0.06528150290250778, 0.11728066205978394, 0.007064842619001865, 0.0914948359131813, 0.04171615466475487, -0.04578579589724541, 0.015482872724533081, 0.06884773075580597, 0.24068933725357056, 0.04607604444026947, -0.02316495217382908, 0.06611938774585724, 0.09430757164955139, 0.14053693413734436, 0.06571704149246216, 0.20859667658805847, 0.16288529336452484, -0.12021251022815704, 0.1287110149860382, 0.041008297353982925, -0.04636544734239578, -0.1389155089855194, 0.016447359696030617, -0.08483490347862244, 0.04940919950604439, -0.045741159468889236, 0.16796143352985382, 0.1276921182870865, -0.18214468657970428, 0.03530742973089218, -0.056089136749506, -0.09714926779270172, -0.08393153548240662, 0.038986846804618835, -0.10069984197616577, -0.1878328174352646, 0.04543133080005646, -0.11444694548845291, 0.019090332090854645, 0.1475221812725067, 0.004217742010951042, 0.007263923995196819, 0.2591799199581146, 0.0358944796025753, 0.05383116006851196, 0.03868423402309418, -0.005068841390311718, -0.025045569986104965, -0.07830184698104858, -0.08966861665248871, 0.03475799411535263, -0.037895750254392624, 0.01900608278810978, -0.07272717356681824, -0.09224291890859604, 0.03233582526445389, 0.025153234601020813, -0.0883057564496994, 0.026987342163920403, 0.023922959342598915, 0.062220700085163116, -0.03813815489411354, 0.00040605448884889483, 0.020512044429779053, -0.019670162349939346, 0.2354629933834076, -0.10205873101949692, -0.10953682661056519, -0.10592455416917801, 0.2169531285762787, 0.052028123289346695, 0.0387786403298378, 0.003583336714655161, -0.1054384857416153, -0.0004870711709372699, 0.204978808760643, 0.11537540704011917, -0.10525877773761749, 0.0012972515542060137, -0.025561459362506866, -0.009979793801903725, -0.06846196949481964, 0.11388427019119263, 0.11985659599304199, 0.026942286640405655, -0.10045027732849121, -0.056537091732025146, -0.026056384667754173, -0.022037846967577934, -0.0314578115940094, 0.022695954889059067, 0.04836532473564148, 0.0396016389131546, -0.0744195505976677, 0.07256808131933212, -0.012481395155191422, -0.1288052499294281, 0.12346100062131882, -0.20221774280071259, -0.12849856913089752, 0.009423313662409782, 0.13337209820747375, -0.02666044794023037, 0.05456521734595299, -0.04073920100927353, -0.02125585824251175, 0.027521196752786636, -0.01426084153354168, -0.0681455209851265, -0.12535661458969116, 0.060489948838949203, -0.13023757934570312, 0.2180333137512207, -0.052831199020147324, 0.08226404339075089, 0.09847927838563919, 0.04237225279211998, -0.046731460839509964, 0.06422773003578186, 0.024076716974377632, -0.13325552642345428, -0.016462987288832664, 0.16275142133235931, -0.057867106050252914, 0.076764777302742, 0.03277008607983589, -0.18134501576423645, 0.020912665873765945, -0.052393775433301926, -0.04894675314426422, -0.036541085690259933, -0.06821262091398239, -0.07680145651102066, 0.10540131479501724, 0.18226373195648193, -0.01028245035558939, 0.051236920058727264, -0.07651548832654953, 0.045449934899806976, 0.0713278278708458, 0.04226996377110481, -0.06343794614076614, -0.26721158623695374, 0.018670925870537758, 0.1125461533665657, -0.055830396711826324, -0.15825198590755463, -0.11270730197429657, 0.017151212319731712, -0.05169486999511719, -0.05799943581223488, 0.08661989122629166, 0.11652053892612457, 0.043318089097738266, -0.062485311180353165, -0.18225592374801636, -0.057628631591796875, 0.17902210354804993, -0.13308124244213104, -0.10899312794208527 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"license": "apache-2.0", "library_name": "transformers"}
text-generation
yam-peleg/Experiment2-7B
[ "transformers", "safetensors", "mistral", "text-generation", "arxiv:1910.09700", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T15:26:05+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #mistral #text-generation #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #mistral #text-generation #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 64, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #mistral #text-generation #arxiv-1910.09700 #license-apache-2.0 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.04166862368583679, 0.19294528663158417, -0.00565074710175395, 0.015576343052089214, 0.09740261733531952, 0.0018807778833433986, 0.05789901316165924, 0.11420097202062607, -0.05003552511334419, 0.12885801494121552, 0.04070472717285156, 0.10962796211242676, 0.11936872452497482, 0.1407015174627304, -0.003504571970552206, -0.2155151218175888, 0.04980916157364845, -0.1058453768491745, -0.01258739922195673, 0.12501691281795502, 0.14908315241336823, -0.0954088643193245, 0.06983769685029984, -0.03609218820929527, -0.016073228791356087, -0.0402071587741375, -0.060646165162324905, -0.041513413190841675, 0.03950463607907295, 0.05431625247001648, 0.06240662559866905, -0.003300471929833293, 0.0801728293299675, -0.28367486596107483, 0.018697958439588547, 0.068385049700737, -0.004608760587871075, 0.0669771134853363, 0.07414057850837708, -0.06557131558656693, 0.11095897853374481, -0.0506414920091629, 0.13285642862319946, 0.08302197605371475, -0.08816267549991608, -0.18223534524440765, -0.09298384934663773, 0.10537559539079666, 0.17730115354061127, 0.05066846311092377, -0.026588909327983856, 0.1006714329123497, -0.07925247400999069, 0.019142232835292816, 0.05144968256354332, -0.09180467575788498, -0.05706454813480377, 0.06526545435190201, 0.09161918610334396, 0.04825152829289436, -0.12598907947540283, -0.034589704126119614, 0.0056123086251318455, 0.017102006822824478, 0.07735797017812729, 0.02069764770567417, 0.14731670916080475, 0.032388463616371155, -0.1297014057636261, -0.05927467346191406, 0.11382096260786057, 0.04015207290649414, -0.04215293377637863, -0.23512817919254303, -0.028819024562835693, -0.012222301214933395, -0.0335053876042366, -0.04219571501016617, 0.04514668136835098, -0.00047637184616178274, 0.09008979052305222, -0.005935803987085819, -0.07398265600204468, -0.03516154736280441, 0.07050351798534393, 0.06780761480331421, 0.030059244483709335, -0.017682623118162155, 0.01944611966609955, 0.10685396194458008, 0.08626683801412582, -0.11604661494493484, -0.05886159837245941, -0.061156801879405975, -0.07161098718643188, -0.03757895156741142, 0.03350892663002014, 0.009119030088186264, 0.07462359964847565, 0.26856333017349243, 0.025587188079953194, 0.05603324621915817, 0.028831996023654938, 0.007935237139463425, 0.04739249870181084, 0.1089349240064621, -0.05712846666574478, -0.12107627838850021, -0.016649138182401657, 0.08437944948673248, 0.026536496356129646, -0.034760136157274246, -0.0417010560631752, 0.06615065038204193, 0.043911732733249664, 0.10984919220209122, 0.10509885102510452, 0.01961970515549183, -0.07238775491714478, -0.05639233440160751, 0.2077396810054779, -0.15489517152309418, 0.03516183793544769, 0.041798185557127, -0.033149976283311844, -0.031306467950344086, 0.01065225712954998, 0.027013929560780525, -0.036672815680503845, 0.09137409180402756, -0.05217616632580757, -0.04674157127737999, -0.10597363114356995, -0.026137787848711014, 0.04449894279241562, 0.01330206636339426, -0.03177689388394356, -0.03566145896911621, -0.07436588406562805, -0.08561325818300247, 0.0869387835264206, -0.06874050199985504, -0.061001889407634735, -0.02138013206422329, -0.0801917016506195, 0.024452297016978264, 0.020871777087450027, 0.07397470623254776, -0.02867235243320465, 0.05468742176890373, -0.05106163024902344, 0.047729142010211945, 0.09779036790132523, 0.035132162272930145, -0.06360576301813126, 0.06066432222723961, -0.22638776898384094, 0.08019262552261353, -0.07270147651433945, 0.06123112142086029, -0.15971983969211578, -0.022097192704677582, 0.0380970723927021, -0.00016348484496120363, -0.007022143341600895, 0.12866158783435822, -0.20674647390842438, -0.019994715228676796, 0.16367171704769135, -0.09709451347589493, -0.07044951617717743, 0.051757436245679855, -0.04413704574108124, 0.09147600084543228, 0.03271377459168434, 0.007501041051000357, 0.06048250198364258, -0.10899953544139862, -0.01165435928851366, -0.05416279658675194, -0.022643128409981728, 0.1340159773826599, 0.08405142277479172, -0.08656053990125656, 0.05779659375548363, 0.02399751916527748, -0.035656314343214035, -0.06690946966409683, -0.014418769627809525, -0.09940238296985626, 0.012407245114445686, -0.06733950972557068, 0.0076343161053955555, -0.018664605915546417, -0.09440974146127701, -0.02771013416349888, -0.1666058897972107, -0.035171132534742355, 0.08134862780570984, -0.0017217934364452958, -0.011632692068815231, -0.10366461426019669, 0.030362889170646667, 0.030370105057954788, 0.0026836544275283813, -0.13047929108142853, -0.03678955137729645, 0.037079811096191406, -0.1558406800031662, 0.03289131820201874, -0.07873660326004028, 0.04977169632911682, 0.014166749082505703, -0.028078405186533928, -0.020859479904174805, 0.017449064180254936, 0.0081904586404562, -0.019382858648896217, -0.22899925708770752, -0.02802218683063984, -0.029544061049818993, 0.1536172777414322, -0.20197926461696625, 0.03410933539271355, 0.07969262450933456, 0.15604744851589203, 0.0032435341272503138, -0.05515560135245323, 0.021976834163069725, -0.06971362978219986, -0.024302059784531593, -0.05630815401673317, 0.0012626007664948702, -0.016396380960941315, -0.04177733138203621, 0.027377402409911156, -0.17498749494552612, -0.04169414937496185, 0.09317784756422043, 0.054987117648124695, -0.11682054400444031, -0.020362254232168198, -0.035645753145217896, -0.05360947921872139, -0.04377356544137001, -0.060842279344797134, 0.10024452209472656, 0.06301113218069077, 0.036907803267240524, -0.0635407343506813, -0.08221858739852905, -0.006284703034907579, -0.017618978396058083, -0.021061228588223457, 0.09222229570150375, 0.07425516098737717, -0.11976317316293716, 0.093970388174057, 0.0874660313129425, 0.06785876303911209, 0.07999815791845322, -0.020717477425932884, -0.07391763478517532, -0.03532349690794945, 0.039611946791410446, 0.019068529829382896, 0.12382332980632782, -0.04680028185248375, 0.04220081865787506, 0.043012309819459915, -0.029560601338744164, 0.017175767570734024, -0.0767202228307724, 0.03359975665807724, 0.020551683381199837, -0.020427212119102478, 0.04948453605175018, -0.037184737622737885, 0.016594747081398964, 0.08402633666992188, 0.058533769100904465, 0.036415163427591324, 0.015351390466094017, -0.05248570069670677, -0.1128775030374527, 0.15880654752254486, -0.11780662089586258, -0.21363064646720886, -0.1330506056547165, 0.024982484057545662, 0.025063807144761086, -0.014864746481180191, 0.005824650637805462, -0.05393596738576889, -0.10789380967617035, -0.09249863773584366, 0.0062092081643640995, 0.05673683062195778, -0.08668006211519241, -0.059869926422834396, 0.04306313395500183, 0.04495549574494362, -0.1424700766801834, 0.020527062937617302, 0.04181644320487976, -0.09161464869976044, -0.015357202850282192, 0.08270744979381561, 0.08016885071992874, 0.18158842623233795, 0.021127747371792793, -0.020351801067590714, 0.028320645913481712, 0.22175416350364685, -0.13565470278263092, 0.11563291400671005, 0.13279883563518524, -0.08048909902572632, 0.08512727916240692, 0.21140246093273163, 0.042638279497623444, -0.09401611983776093, 0.028545530512928963, 0.03357614949345589, -0.02403010055422783, -0.23939213156700134, -0.07092683017253876, -0.0013685966841876507, -0.06716125458478928, 0.07811819761991501, 0.09883560985326767, 0.0776619166135788, 0.0210383590310812, -0.09727127104997635, -0.09041786193847656, 0.05844145268201828, 0.11003929376602173, 0.005977734923362732, -0.0010036816820502281, 0.08619128912687302, -0.03526197373867035, 0.02053396962583065, 0.08993267267942429, 0.012363693676888943, 0.1520329713821411, 0.047393251210451126, 0.17737804353237152, 0.0840906947851181, 0.07860663533210754, -0.0004794647975359112, 0.006364364642649889, 0.012932327575981617, 0.04642070084810257, -0.006052643060684204, -0.08458072692155838, -0.027158472687005997, 0.11165141314268112, 0.06500331312417984, 0.015393076464533806, 0.020406542345881462, -0.05238749086856842, 0.08462364226579666, 0.19093233346939087, -0.006165898405015469, -0.1801624298095703, -0.059130482375621796, 0.07549434900283813, -0.0990021824836731, -0.10064712166786194, -0.0039864154532551765, 0.014100136235356331, -0.16932961344718933, 0.04136020317673683, -0.02567523531615734, 0.10914346575737, -0.1284799426794052, -0.02066126838326454, 0.079505056142807, 0.06859999150037766, -0.0012688254937529564, 0.060875728726387024, -0.18528470396995544, 0.09756795316934586, 0.010917199775576591, 0.06973090022802353, -0.09255387634038925, 0.0928410217165947, -0.00668302970007062, -0.027202703058719635, 0.14476221799850464, -0.001775130513124168, -0.07416173070669174, -0.05728907883167267, -0.09669062495231628, -0.008932547643780708, 0.11787547916173935, -0.133856400847435, 0.08551253378391266, -0.032557401806116104, -0.03564809262752533, -0.013994505628943443, -0.08327500522136688, -0.1109219491481781, -0.1709768921136856, 0.059307605028152466, -0.12648512423038483, 0.04020201787352562, -0.1088717058300972, -0.02373320981860161, -0.027199482545256615, 0.1699579954147339, -0.2393503487110138, -0.0769786387681961, -0.14049221575260162, -0.10581114888191223, 0.12965087592601776, -0.05028373748064041, 0.09073053300380707, -0.022501198574900627, 0.15729914605617523, 0.01874421164393425, -0.021332228556275368, 0.08108112961053848, -0.08612661808729172, -0.1987118273973465, -0.06719952821731567, 0.16559822857379913, 0.11229605972766876, 0.031270451843738556, -0.0012020005378872156, 0.03954574465751648, -0.025526942685246468, -0.11973368376493454, 0.021365778520703316, 0.15028510987758636, 0.06962436437606812, 0.007621194235980511, -0.016045305877923965, -0.11842469125986099, -0.07784009724855423, -0.028162069618701935, 0.023731907829642296, 0.16045090556144714, -0.07187303155660629, 0.17342956364154816, 0.1463107019662857, -0.059301216155290604, -0.2025192379951477, -0.0072204358875751495, 0.02655131369829178, -0.015131231397390366, 0.009906691499054432, -0.18563494086265564, 0.08842182159423828, 0.0035971112083643675, -0.057965271174907684, 0.09906121343374252, -0.16108983755111694, -0.1368165910243988, 0.08425280451774597, 0.0501166433095932, -0.19157421588897705, -0.139436736702919, -0.10083521902561188, -0.043168213218450546, -0.16376076638698578, 0.09043843299150467, 0.01753687486052513, 0.010611959733068943, 0.027408726513385773, 0.012237385846674442, 0.02259771153330803, -0.049664974212646484, 0.17527315020561218, -0.0119782704859972, 0.024203931912779808, -0.09571193903684616, -0.08417301625013351, 0.01689862087368965, -0.05036649480462074, 0.07465502619743347, -0.02852136269211769, 0.0146928196772933, -0.10245449095964432, -0.03361695632338524, -0.046283259987831116, 0.018411923199892044, -0.0984109491109848, -0.08554413914680481, -0.052167847752571106, 0.08726155012845993, 0.09808032214641571, -0.020503507927060127, -0.018636612221598625, -0.07416849583387375, 0.05757380276918411, 0.2149011194705963, 0.18108037114143372, 0.04631878063082695, -0.07480046898126602, -0.004399713594466448, -0.015207556076347828, 0.04487600550055504, -0.19843150675296783, 0.05744349583983421, 0.05550002306699753, 0.02062990516424179, 0.10227029025554657, -0.024344047531485558, -0.15487264096736908, -0.07267282158136368, 0.06276534497737885, -0.05848631262779236, -0.20858339965343475, 0.010548625141382217, 0.05569260194897652, -0.17460303008556366, -0.034738194197416306, 0.0456136129796505, -0.007365865167230368, -0.03797522932291031, 0.020451541990041733, 0.09710922092199326, 0.0038564593996852636, 0.08027420938014984, 0.07102498412132263, 0.08460576832294464, -0.09778829663991928, 0.09052757918834686, 0.09921758621931076, -0.06244191899895668, 0.02659420855343342, 0.09714852273464203, -0.05697975680232048, -0.03690675273537636, 0.038184426724910736, 0.07610335201025009, 0.027226708829402924, -0.04769636318087578, 0.008859969675540924, -0.0913708433508873, 0.06549783051013947, 0.10440699011087418, 0.03000110760331154, 0.02052699401974678, 0.04642310366034508, 0.04275054112076759, -0.06684256345033646, 0.12171297520399094, 0.03287801519036293, 0.014797203242778778, -0.041677236557006836, -0.046708397567272186, 0.010782824829220772, -0.031146129593253136, -0.003426467766985297, -0.0212049949914217, -0.08137737214565277, -0.015304007567465305, -0.13043250143527985, 0.00355430762283504, -0.06720879673957825, 0.015176482498645782, 0.023503823205828667, -0.03384915739297867, 0.008213633671402931, 0.009011444635689259, -0.06849221140146255, -0.06852424889802933, -0.013598221354186535, 0.09843763709068298, -0.16962307691574097, 0.029034918174147606, 0.08575760573148727, -0.10844960063695908, 0.10187135636806488, 0.008888037875294685, -0.009416608139872551, 0.018001845106482506, -0.15660931169986725, 0.04044801741838455, -0.037415020167827606, 0.006806433200836182, 0.015853602439165115, -0.20005734264850616, -0.0019246236188337207, -0.03177458792924881, -0.0705052837729454, -0.010842126794159412, -0.016560347750782967, -0.1186550036072731, 0.10135795176029205, 0.004299563821405172, -0.08060503005981445, -0.029897188767790794, 0.030650708824396133, 0.07598836719989777, -0.031478025019168854, 0.15097710490226746, -0.011336207389831543, 0.06422024965286255, -0.1609204262495041, -0.010663383640348911, -0.008957091718912125, 0.01420842669904232, -0.05656726285815239, -0.001103369751945138, 0.04814773052930832, -0.014907282777130604, 0.17374174296855927, -0.034365665167570114, 0.011136728338897228, 0.06490659713745117, 0.058584485203027725, -0.027248801663517952, 0.0942847952246666, 0.04749126732349396, 0.014289948157966137, 0.007745350245386362, 0.01487020868808031, -0.047270435839891434, -0.03966875746846199, -0.19174465537071228, 0.06610973924398422, 0.19794288277626038, 0.1044018343091011, -0.020746521651744843, 0.06986040621995926, -0.10006950795650482, -0.10040159523487091, 0.14918941259384155, -0.03457310050725937, -0.0025222725234925747, -0.07169237732887268, 0.12801261246204376, 0.14952176809310913, -0.1830597221851349, 0.06886568665504456, -0.06775565445423126, -0.03977802023291588, -0.10651897639036179, -0.201371967792511, -0.06249268725514412, -0.04581226781010628, -0.017517665401101112, -0.04613880068063736, 0.06678374856710434, 0.07430177181959152, -0.006824250798672438, -0.007840139791369438, 0.0655519962310791, -0.036141421645879745, -0.0053302873857319355, 0.027680065482854843, 0.059438642114400864, 0.008952193893492222, -0.033686328679323196, 0.015949474647641182, -0.010523517616093159, 0.05258147791028023, 0.07987221330404282, 0.05156650394201279, -0.01909230649471283, 0.021411675959825516, -0.03876841068267822, -0.1029580757021904, 0.05319680646061897, -0.02604341320693493, -0.07099205255508423, 0.15270604193210602, 0.021440722048282623, 0.007952463813126087, -0.007006566505879164, 0.2409990429878235, -0.06405144929885864, -0.10283639281988144, -0.14431513845920563, 0.07044614851474762, -0.04318870231509209, 0.04597603902220726, 0.0419544093310833, -0.11124377697706223, 0.026897640898823738, 0.14373010396957397, 0.1525527536869049, -0.028645912185311317, 0.021028004586696625, 0.031088391318917274, 0.007085015065968037, -0.020426327362656593, 0.03804256394505501, 0.0569956935942173, 0.1498127281665802, -0.049512092024087906, 0.07898244261741638, 0.00368340197019279, -0.08552169054746628, -0.03570893406867981, 0.11698101460933685, -0.021283045411109924, 0.007356108166277409, -0.058085665106773376, 0.12010903656482697, -0.06618686020374298, -0.21936537325382233, 0.038884084671735764, -0.06754741072654724, -0.1315430998802185, -0.02041028067469597, 0.07517372071743011, -0.008638354949653149, 0.019841624423861504, 0.08050349354743958, -0.07101814448833466, 0.1898367553949356, 0.03590880706906319, -0.06227270886301994, -0.05171479657292366, 0.07330481708049774, -0.07958567887544632, 0.29808610677719116, 0.016964634880423546, 0.04131867364048958, 0.10863476991653442, -0.012988881208002567, -0.1398736834526062, 0.029780730605125427, 0.09792774170637131, -0.09334233403205872, 0.05595870316028595, 0.17345324158668518, 0.0029040013905614614, 0.1337554007768631, 0.07441878318786621, -0.07816100865602493, 0.04427627474069595, -0.0647587776184082, -0.07012900710105896, -0.10388600081205368, 0.1026725023984909, -0.09383752197027206, 0.14164794981479645, 0.11840517818927765, -0.05714124068617821, 0.007326686754822731, -0.03666400909423828, 0.04674949124455452, -0.005353722721338272, 0.11694536358118057, 0.01294570043683052, -0.18544849753379822, 0.02969195321202278, -0.02853630855679512, 0.10067041218280792, -0.15941902995109558, -0.08449898660182953, 0.04787616431713104, 0.009869850240647793, -0.06761465966701508, 0.12036609649658203, 0.05896257236599922, 0.026718489825725555, -0.04979591816663742, -0.03311346471309662, -0.01145645696669817, 0.1395922303199768, -0.1021265834569931, -0.005856354255229235 ]
null
null
peft
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.8.2
{"library_name": "peft", "base_model": "bigscience/bloom-3b"}
null
vmathur87/llm-support
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:bigscience/bloom-3b", "region:us" ]
2024-02-10T15:26:37+00:00
[ "1910.09700" ]
[]
TAGS #peft #safetensors #arxiv-1910.09700 #base_model-bigscience/bloom-3b #region-us
# Model Card for Model ID ## Model Details ### Model Description - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact ### Framework versions - PEFT 0.8.2
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ "TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-bigscience/bloom-3b #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ 35, 6, 3, 54, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4, 11 ]
[ "passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-bigscience/bloom-3b #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2" ]
[ -0.09975722432136536, 0.1893470585346222, -0.0036378027871251106, 0.03377080708742142, 0.08956268429756165, 0.01606217958033085, 0.04934121295809746, 0.12286069244146347, -0.03180153667926788, 0.10665926337242126, 0.07045666873455048, 0.10034925490617752, 0.0986793041229248, 0.21037167310714722, 0.010604910552501678, -0.20116940140724182, 0.02554682455956936, -0.09600938856601715, -0.014991069212555885, 0.1206747442483902, 0.1542518436908722, -0.09912464022636414, 0.0753534585237503, -0.018081873655319214, -0.014627320691943169, -0.03511117398738861, -0.0738551914691925, -0.033506445586681366, 0.04699651524424553, 0.04367610812187195, 0.055165186524391174, -0.0007730853394605219, 0.0822964757680893, -0.2646140158176422, 0.0199875608086586, 0.03787897527217865, -0.0047507029958069324, 0.08792980760335922, 0.09946314990520477, -0.037762485444545746, 0.11771578341722488, -0.04130297526717186, 0.14244452118873596, 0.07968951761722565, -0.09225821495056152, -0.21491482853889465, -0.0703209936618805, 0.08734104037284851, 0.1642523854970932, 0.07540570944547653, -0.0409468412399292, 0.1252894401550293, -0.10191884636878967, 0.01654636673629284, 0.05280008167028427, -0.07238156348466873, -0.06962985545396805, 0.05547565594315529, 0.11157765239477158, 0.06257667392492294, -0.13659654557704926, -0.03524216637015343, 0.017521031200885773, 0.038271091878414154, 0.08032975345849991, 0.0212861355394125, 0.14492926001548767, 0.032858919352293015, -0.14962902665138245, -0.036321334540843964, 0.13016711175441742, 0.030908193439245224, -0.03823449835181236, -0.2261674702167511, 0.009463529102504253, -0.08858253806829453, -0.028639553114771843, -0.05611279234290123, 0.04277198389172554, -0.0017891575116664171, 0.09335581958293915, -0.029236914590001106, -0.0928860753774643, -0.006158793345093727, 0.09257356077432632, 0.049603354185819626, 0.02171463705599308, -0.020311374217271805, 0.004161885008215904, 0.11514034122228622, 0.05869006738066673, -0.12496386468410492, -0.06233423575758934, -0.0693092867732048, -0.04235416650772095, -0.04913373664021492, 0.03177841007709503, 0.05083855614066124, 0.06261835247278214, 0.2525183856487274, -0.028780506923794746, 0.049934059381484985, 0.06158886104822159, 0.01922941766679287, 0.04768817499279976, 0.0908169373869896, -0.059672996401786804, -0.14890699088573456, -0.010012786835432053, 0.09345709532499313, -0.006876375060528517, -0.017138052731752396, -0.046624407172203064, 0.036221787333488464, 0.044156838208436966, 0.10888443887233734, 0.0919705480337143, -0.01565215177834034, -0.0773858055472374, -0.049856461584568024, 0.21394090354442596, -0.14641202986240387, 0.040052153170108795, 0.02451586164534092, -0.017240051180124283, -0.04536481201648712, 0.002308693714439869, 0.018550293520092964, -0.015214066952466965, 0.0972033143043518, -0.06477192044258118, -0.03805689886212349, -0.11374577134847641, -0.016620837152004242, 0.038670241832733154, 0.01875949651002884, -0.02165951579809189, -0.030211161822080612, -0.05726774036884308, -0.09100127965211868, 0.10074631124734879, -0.0685826987028122, -0.06518290191888809, -0.024242497980594635, -0.09513060003519058, 0.01873389631509781, 0.02969704009592533, 0.11462023109197617, -0.02456081658601761, 0.039704736322164536, -0.019116638228297234, 0.06410972028970718, 0.07861677557229996, 0.032008204609155655, -0.07036947458982468, 0.06277956813573837, -0.19494210183620453, 0.08653482794761658, -0.07808511704206467, 0.029986953362822533, -0.1642191857099533, -0.01953001879155636, 0.00724397087469697, 0.017147352918982506, 0.03540012985467911, 0.14984478056430817, -0.19746874272823334, -0.02853182516992092, 0.15303878486156464, -0.09353726357221603, -0.12432079762220383, 0.04768049344420433, -0.05718144401907921, 0.1592889130115509, 0.018705099821090698, -0.011164537630975246, 0.08360128104686737, -0.14598457515239716, -0.02864304929971695, -0.03029758855700493, -0.005292677786201239, 0.10320105403661728, 0.08967240899801254, -0.07198780030012131, 0.027997728437185287, 0.014321189373731613, -0.04353460669517517, -0.028955575078725815, -0.052595168352127075, -0.11167800426483154, 0.0015150081599131227, -0.08271915465593338, 0.024840297177433968, -0.018803343176841736, -0.07312550395727158, -0.01649562269449234, -0.15794962644577026, -0.02461201697587967, 0.08167023211717606, 0.021112259477376938, -0.015340887941420078, -0.08818652480840683, 0.03270765393972397, -0.029429657384753227, -0.02562224306166172, -0.15098778903484344, -0.02566339634358883, 0.02210245095193386, -0.1389470100402832, 0.01246930006891489, -0.11754001677036285, 0.06532931327819824, 0.011274385266005993, -0.06795505434274673, -0.03228362277150154, -0.014710618183016777, 0.0070625836960971355, -0.05275893583893776, -0.23767872154712677, -0.021246425807476044, -0.056716758757829666, 0.14908574521541595, -0.2188778519630432, 0.037142857909202576, 0.042639121413230896, 0.11916908621788025, 0.004382405895739794, -0.05946265906095505, 0.02613980695605278, -0.07412714511156082, -0.02361799031496048, -0.06931019574403763, -0.0068540796637535095, -0.006426749285310507, -0.04797806590795517, 0.014037731103599072, -0.1123562753200531, -0.04244823753833771, 0.09795384854078293, 0.07213319838047028, -0.1659919023513794, -0.02522103674709797, -0.04826221615076065, -0.06394803524017334, -0.0856405720114708, -0.06031101942062378, 0.09818363189697266, 0.049710601568222046, 0.032421864569187164, -0.0732237845659256, -0.06961683183908463, 0.006064711604267359, -0.022524194791913033, -0.02401241473853588, 0.11832126975059509, 0.06971362978219986, -0.1206585168838501, 0.09845659136772156, 0.07528426498174667, 0.027226094156503677, 0.07760791480541229, -0.024279430508613586, -0.10515666007995605, -0.03277328610420227, 0.040167830884456635, 0.012574390508234501, 0.1585487723350525, -0.0718303918838501, 0.05668378621339798, 0.04635933041572571, -0.03491443395614624, 0.048647649586200714, -0.09140434861183167, 0.010275826789438725, 0.00594200287014246, -0.011842416599392891, 0.024223310872912407, -0.023575711995363235, 0.010504592210054398, 0.08822529762983322, 0.057430658489465714, 0.03472631424665451, 0.035188037902116776, -0.02726294845342636, -0.12781551480293274, 0.18317241966724396, -0.0980101004242897, -0.23131372034549713, -0.155201256275177, 0.05162796378135681, 0.057933349162340164, -0.017597876489162445, 0.025013238191604614, -0.053373027592897415, -0.10238201916217804, -0.07824961096048355, 0.0032054490875452757, 0.03306940197944641, -0.058518365025520325, -0.07132046669721603, 0.05410126596689224, 0.03884412348270416, -0.11645375192165375, 0.036966800689697266, 0.05729293078184128, -0.010749425739049911, 0.006462321616709232, 0.058341022580862045, 0.0769619345664978, 0.17465059459209442, -0.009634934365749359, -0.009058217518031597, 0.053868018090724945, 0.2827358543872833, -0.15757958590984344, 0.11535574495792389, 0.12487572431564331, -0.06081799790263176, 0.07559762895107269, 0.19061574339866638, 0.03549840673804283, -0.0989280641078949, 0.03910971060395241, 0.025636713951826096, -0.02745332196354866, -0.2751244902610779, -0.04868477210402489, -0.01140675414353609, -0.09589074552059174, 0.07442373037338257, 0.09076731652021408, 0.08179663121700287, 0.04038264974951744, -0.06491625308990479, -0.09625113010406494, 0.03185315057635307, 0.09958431124687195, -0.028621232137084007, 0.004949393216520548, 0.0845138356089592, -0.02744724042713642, 0.008226574398577213, 0.09415486454963684, -0.010531400330364704, 0.1604742705821991, 0.05617470666766167, 0.11773977428674698, 0.07711881399154663, 0.08479076623916626, -0.0016485498053953052, 0.025294769555330276, 0.016279572620987892, 0.017687035724520683, 0.013126881793141365, -0.08280302584171295, 0.03816579282283783, 0.1121472716331482, 0.04097863286733627, 0.03123302571475506, 0.01522710733115673, -0.042750827968120575, 0.0501626580953598, 0.1826663315296173, 0.011048371903598309, -0.1937234252691269, -0.08176153153181076, 0.06406830996274948, -0.07898042351007462, -0.13738738000392914, -0.019705820828676224, 0.028679000213742256, -0.1673949956893921, 0.008052579127252102, -0.04078042134642601, 0.09715253859758377, -0.061438869684934616, -0.03795260936021805, 0.08920086175203323, 0.07131072133779526, -0.027115363627672195, 0.06199248880147934, -0.20471739768981934, 0.11680953949689865, 0.020494312047958374, 0.06664200127124786, -0.0932881161570549, 0.09758492559194565, -0.0010622412664815784, -0.014568718150258064, 0.16814740002155304, 0.0033798536751419306, -0.07787255942821503, -0.05395779013633728, -0.09767833352088928, -0.012810480780899525, 0.10179094225168228, -0.12889918684959412, 0.06576649844646454, -0.015660583972930908, -0.03250566124916077, 0.005942856427282095, -0.0711740106344223, -0.12040203809738159, -0.1729748249053955, 0.05535253882408142, -0.10745986551046371, 0.03623555600643158, -0.09312895685434341, -0.06493029743432999, 0.005988468416035175, 0.1816624104976654, -0.18587782979011536, -0.09241292625665665, -0.14343905448913574, -0.09036695212125778, 0.1693897694349289, -0.039945751428604126, 0.09023533761501312, 0.004654248710721731, 0.17036733031272888, 0.013007830828428268, -0.001251792535185814, 0.10266374796628952, -0.0921061635017395, -0.20264814794063568, -0.05711900815367699, 0.170995831489563, 0.13730528950691223, 0.046564362943172455, -0.010836083441972733, 0.02817978523671627, -0.05975474417209625, -0.11247469484806061, 0.027004418894648552, 0.13228382170200348, 0.07796267420053482, -0.009910098277032375, -0.03963500261306763, -0.09661024063825607, -0.0610857829451561, -0.058929990977048874, 0.017426129430532455, 0.19666948914527893, -0.07486938685178757, 0.17224501073360443, 0.11764737963676453, -0.0544845387339592, -0.20620942115783691, 0.053339678794145584, 0.05797739326953888, 0.015108970925211906, 0.03403879329562187, -0.19606147706508636, 0.0988064557313919, 0.009693064726889133, -0.07434016466140747, 0.15721504390239716, -0.16550388932228088, -0.14543704688549042, 0.09085861593484879, 0.031258851289749146, -0.22248686850070953, -0.14204736053943634, -0.0962815210223198, -0.02642999403178692, -0.1015612781047821, 0.0755636990070343, 0.011053462512791157, 0.014060859568417072, 0.026317108422517776, 0.020976373925805092, 0.027855530381202698, -0.05372444540262222, 0.20438621938228607, -0.026720819994807243, 0.008135526441037655, -0.04723326116800308, -0.09648061543703079, 0.03731009364128113, -0.048261236399412155, 0.10010926425457001, 0.00132300378754735, 0.02478194236755371, -0.13765224814414978, -0.04189179092645645, -0.0641426220536232, 0.03166606277227402, -0.09901221096515656, -0.09189143031835556, -0.051526155322790146, 0.09958703070878983, 0.0983964279294014, -0.032119471579790115, 0.009831064380705357, -0.08359651267528534, 0.06998662650585175, 0.19518674910068512, 0.18952326476573944, 0.06973535567522049, -0.06460649520158768, 0.02517933025956154, -0.03528497368097305, 0.03818199411034584, -0.2242344617843628, 0.04508070647716522, 0.05750501528382301, 0.02269863523542881, 0.0858859121799469, -0.007428336422890425, -0.1496102660894394, -0.07465948909521103, 0.0798991322517395, -0.04438355192542076, -0.16907872259616852, -0.02404472976922989, 0.054965659976005554, -0.21798112988471985, -0.04687943309545517, 0.015718305483460426, -0.01840486004948616, -0.04578475281596184, 0.02102261781692505, 0.0858994573354721, -0.019547034054994583, 0.1108282133936882, 0.09664713591337204, 0.09561387449502945, -0.09917635470628738, 0.07368329167366028, 0.07448484748601913, -0.046363599598407745, 0.02335950918495655, 0.11153175681829453, -0.04270366579294205, -0.03507141396403313, 0.09748817980289459, 0.07705319672822952, 0.02587110735476017, -0.04290695860981941, 0.012588207609951496, -0.06199553236365318, 0.06379231810569763, 0.11365632712841034, 0.030371256172657013, -0.00734240235760808, 0.05562450364232063, 0.0342625267803669, -0.10192793607711792, 0.11105329543352127, 0.05220925435423851, 0.02622404880821705, -0.0381653755903244, -0.02980240248143673, -0.018469011411070824, -0.009976601228117943, -0.01961338333785534, -0.0014100528787821531, -0.09227113425731659, -0.011602005921304226, -0.10153832286596298, 0.031485896557569504, -0.07173488289117813, 0.01048267725855112, 0.023646535351872444, -0.04774729907512665, 0.009758591651916504, 0.00464885588735342, -0.0776129961013794, -0.05043148621916771, -0.01859482191503048, 0.08593504130840302, -0.12590427696704865, 0.031011343002319336, 0.07912012934684753, -0.10972778499126434, 0.07716438174247742, 0.0008825258701108396, 0.011078757233917713, 0.006712749600410461, -0.16461843252182007, 0.05249214172363281, -0.01905198208987713, -0.011518056504428387, 0.01272913534194231, -0.21872104704380035, -0.014998218975961208, -0.04613448679447174, -0.050736501812934875, 0.011610143817961216, -0.033190932124853134, -0.12563230097293854, 0.10091212391853333, -0.010568798519670963, -0.06953324377536774, -0.018838394433259964, 0.030913589522242546, 0.08951631188392639, -0.022439107298851013, 0.1290925294160843, -0.02153109200298786, 0.07500520348548889, -0.17019806802272797, -0.0026300798635929823, -0.020199008285999298, 0.03451049327850342, -0.029923763126134872, -0.018896542489528656, 0.057657066732645035, -0.015780624002218246, 0.1843862235546112, -0.031184332445263863, 0.06503761559724808, 0.05767204985022545, 0.013061377219855785, 0.015329992398619652, 0.08940810710191727, 0.06578909605741501, -0.0037376252003014088, -0.006552417762577534, 0.0407535620033741, -0.009051628410816193, -0.03856135159730911, -0.15013954043388367, 0.07544232159852982, 0.16951948404312134, 0.04490186646580696, 0.0165391955524683, 0.02919512242078781, -0.11594906449317932, -0.06822231411933899, 0.12260214239358902, -0.0047270748764276505, -0.04092750325798988, -0.07362427562475204, 0.18552640080451965, 0.12788769602775574, -0.20165511965751648, 0.07690855115652084, -0.062116991728544235, -0.054208334535360336, -0.12813161313533783, -0.14451885223388672, -0.06660336256027222, -0.03386666998267174, -0.02451177127659321, -0.06747842580080032, 0.04866989701986313, 0.05827957019209862, 0.008104433305561543, -0.018369058147072792, 0.09989504516124725, 0.01260843500494957, -0.021809840574860573, 0.0482124388217926, 0.06341245025396347, 0.028432775288820267, -0.0963190421462059, 0.008414185605943203, -0.002993077039718628, 0.019331194460392, 0.06435688585042953, 0.019559256732463837, -0.05534745380282402, 0.005792077165096998, -0.019682414829730988, -0.1194445937871933, 0.04103434458374977, -0.02048632688820362, -0.03595961630344391, 0.13712374866008759, 0.026314856484532356, 0.008617603220045567, -0.021309740841388702, 0.24031347036361694, -0.07093461602926254, -0.07598961889743805, -0.14398452639579773, 0.06201057881116867, -0.07021069526672363, 0.032868802547454834, 0.0316629633307457, -0.11284003406763077, 0.017592448741197586, 0.15055249631404877, 0.13176535069942474, -0.015685388818383217, 0.01186107937246561, 0.05260205268859863, 0.002323627471923828, -0.032105494290590286, 0.019202426075935364, 0.056421153247356415, 0.14102327823638916, -0.07660733908414841, 0.07320397347211838, -0.009290932677686214, -0.07478061318397522, -0.024547984823584557, 0.11028088629245758, -0.0019461424089968204, 0.002267061732709408, -0.0683116614818573, 0.13425426185131073, -0.09255504608154297, -0.23336157202720642, 0.05031728371977806, -0.0727153867483139, -0.1565227508544922, -0.05105594918131828, 0.016761399805545807, -0.016656961292028427, 0.017976684495806694, 0.08014219254255295, -0.04792019724845886, 0.16093593835830688, 0.04630555585026741, -0.05380043759942055, -0.07456637173891068, 0.06712479144334793, -0.12380830198526382, 0.27879318594932556, 0.024334311485290527, 0.05216830596327782, 0.10606543719768524, -0.018513860180974007, -0.14877784252166748, 0.00756117282435298, 0.10665912181138992, -0.06999801844358444, 0.06708445399999619, 0.177728533744812, 0.0010692300274968147, 0.12283014506101608, 0.05960458144545555, -0.059786297380924225, 0.033235616981983185, -0.0815848633646965, -0.05117927864193916, -0.1116955503821373, 0.07964935153722763, -0.08103553205728531, 0.15869086980819702, 0.1284148395061493, -0.07295882701873779, -0.006487618200480938, -0.02283477783203125, 0.08460917323827744, 0.013708834536373615, 0.10979172587394714, 0.006446777377277613, -0.19367627799510956, 0.03685363754630089, 0.010964524932205677, 0.10468564182519913, -0.20935799181461334, -0.06921717524528503, 0.0484384149312973, -0.014200243167579174, -0.08089397847652435, 0.11299072206020355, 0.04689827188849449, 0.031133469194173813, -0.03836746886372566, -0.03776996210217476, 0.0004732717643491924, 0.14430396258831024, -0.11418819427490234, -0.007164023816585541 ]
null
null
transformers
Copied from https://huggingface.co/susnato/phi-2 commit@9070ddb4fce238899ddbd2aca1faf6a0aeb6e444. Temporarily used for [Bunny](https://github.com/BAAI-DCAI/Bunny). ---------------------------------------------------- **DISCLAIMER**: I don't own the weights to this model, this is a property of Microsoft and taken from their official repository : [microsoft/phi-2](https://huggingface.co/microsoft/phi-2). The sole purpose of this repository is to use this model through the `transformers` API or to load and use the model using the HuggingFace `transformers` library. # Usage First make sure you have the latest version of the `transformers` installed. ``` pip install -U transformers ``` Then use the transformers library to load the model from the library itself ```python from transformers import AutoModelForCausalLM, AutoTokenizer model = AutoModelForCausalLM.from_pretrained("susnato/phi-2") tokenizer = AutoTokenizer.from_pretrained("susnato/phi-2") inputs = tokenizer('''def print_prime(n): """ Print all primes between 1 and n """''', return_tensors="pt", return_attention_mask=False) outputs = model.generate(**inputs, max_length=200) text = tokenizer.batch_decode(outputs)[0] print(text) ```
{}
text-generation
Isaachhe/phi-2_dev
[ "transformers", "pytorch", "phi", "text-generation", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-10T15:27:09+00:00
[]
[]
TAGS #transformers #pytorch #phi #text-generation #autotrain_compatible #endpoints_compatible #region-us
Copied from URL commit@9070ddb4fce238899ddbd2aca1faf6a0aeb6e444. Temporarily used for Bunny. ---------------------------------------------------- DISCLAIMER: I don't own the weights to this model, this is a property of Microsoft and taken from their official repository : microsoft/phi-2. The sole purpose of this repository is to use this model through the 'transformers' API or to load and use the model using the HuggingFace 'transformers' library. # Usage First make sure you have the latest version of the 'transformers' installed. Then use the transformers library to load the model from the library itself
[ "# Usage \n\nFirst make sure you have the latest version of the 'transformers' installed.\n\n\n\nThen use the transformers library to load the model from the library itself" ]
[ "TAGS\n#transformers #pytorch #phi #text-generation #autotrain_compatible #endpoints_compatible #region-us \n", "# Usage \n\nFirst make sure you have the latest version of the 'transformers' installed.\n\n\n\nThen use the transformers library to load the model from the library itself" ]
[ 36, 36 ]
[ "passage: TAGS\n#transformers #pytorch #phi #text-generation #autotrain_compatible #endpoints_compatible #region-us \n# Usage \n\nFirst make sure you have the latest version of the 'transformers' installed.\n\n\n\nThen use the transformers library to load the model from the library itself" ]
[ -0.050090741366147995, 0.04286512732505798, -0.003643230302259326, -0.026060571894049644, 0.20317041873931885, 0.030670875683426857, 0.094554603099823, 0.03822524473071098, -0.016184037551283836, -0.04122554883360863, 0.1111297458410263, 0.21943235397338867, -0.007515890058130026, 0.047256965190172195, -0.03464386239647865, -0.22980454564094543, 0.057872068136930466, 0.14483240246772766, -0.164666086435318, 0.10256649553775787, 0.06599471718072891, -0.08203897625207901, 0.09636104106903076, 0.021983826532959938, -0.18914353847503662, 0.008082173764705658, 0.027367550879716873, -0.12151124328374863, 0.11500662565231323, 0.0432121567428112, 0.1464085727930069, 0.03605373948812485, 0.026048358529806137, -0.05221360921859741, 0.0313156358897686, 0.056542158126831055, -0.009518629871308804, 0.06450929492712021, 0.04209575057029724, -0.08426222950220108, 0.180099755525589, 0.049948129802942276, 0.06408462673425674, 0.019765082746744156, -0.08547268062829971, -0.13304898142814636, -0.0063406131230294704, 0.08042769134044647, 0.14688245952129364, 0.12427352368831635, 0.03566470742225647, 0.2052491158246994, 0.0031495497096329927, 0.11147656291723251, 0.0943172499537468, -0.19459815323352814, 0.009721212089061737, 0.07162642478942871, 0.08698830753564835, 0.03489150106906891, -0.015881245955824852, 0.053242117166519165, 0.04192550107836723, 0.04130249097943306, 0.1034000888466835, -0.06963488459587097, -0.029387762770056725, -0.027150392532348633, -0.1376282125711441, -0.05404297262430191, 0.17249174416065216, -0.06348173320293427, 0.026899760589003563, -0.0045479112304747105, -0.15038041770458221, 0.031695377081632614, -0.014448471367359161, -0.020230110734701157, -0.016361599788069725, 0.009136664681136608, -0.10032759606838226, -0.1558368057012558, -0.08228675276041031, -0.07724035531282425, -0.16783900558948517, 0.25476640462875366, 0.020075177773833275, 0.05568497255444527, -0.12900257110595703, 0.06224588304758072, 0.04031611606478691, -0.10720806568861008, -0.013608190231025219, -0.07551709562540054, 0.011081415228545666, -0.056999482214450836, -0.08698532730340958, -0.0989188626408577, 0.02229917049407959, 0.18893718719482422, 0.02737937495112419, -0.028889022767543793, -0.02903931960463524, 0.03582410514354706, 0.05647922307252884, 0.1556922346353531, -0.013675040565431118, 0.0677105113863945, 0.08591621369123459, -0.07033275812864304, 0.0008663515327498317, -0.04612915217876434, -0.10679397732019424, -0.018082771450281143, 0.045299090445041656, 0.07769429683685303, 0.06180110573768616, 0.06996060162782669, 0.10184977203607559, -0.08843759447336197, 0.11915478855371475, -0.09742750972509384, -0.016116511076688766, -0.014990976080298424, -0.017495891079306602, 0.13670413196086884, 0.1107562705874443, -0.020244557410478592, -0.07311573624610901, 0.07112245261669159, -0.06486278027296066, -0.01391420979052782, -0.07261496782302856, -0.09111764281988144, 0.01747412607073784, -0.10489799827337265, 0.06730538606643677, -0.22807447612285614, -0.24333587288856506, 0.013697736896574497, 0.055680111050605774, 0.017335960641503334, -0.042652443051338196, 0.08336939662694931, 0.008430157788097858, -0.017977062612771988, -0.07289153337478638, -0.01039864867925644, -0.041590265929698944, 0.11266981810331345, 0.09580434113740921, 0.027495553717017174, -0.07531245052814484, 0.08488812297582626, -0.14009183645248413, -0.015608450397849083, -0.16315774619579315, 0.06860143691301346, -0.023631006479263306, 0.07495373487472534, -0.081601083278656, -0.039595507085323334, -0.010563417337834835, 0.04360455274581909, 0.006418750621378422, 0.12289934605360031, -0.0227018091827631, -0.12532009184360504, 0.12093938142061234, -0.20370563864707947, -0.18810337781906128, 0.04927675053477287, 0.004680545534938574, 0.022449631243944168, 0.09498392045497894, 0.01035273727029562, 0.19954830408096313, -0.09628010541200638, 0.046742625534534454, 0.11052370071411133, -0.09526993334293365, -0.14173972606658936, 0.02332426980137825, 0.06699872761964798, -0.05592052638530731, 0.001586963888257742, -0.15429599583148956, 0.18633095920085907, -0.017816618084907532, -0.04822934791445732, -0.09172586351633072, -0.10870671272277832, 0.053774721920490265, 0.07210239768028259, 0.09313258528709412, 0.06056110933423042, 0.0020505792926996946, 0.24797619879245758, 0.10139820724725723, -0.04639474302530289, 0.000058123154303757474, -0.05843479931354523, 0.09860116243362427, -0.04828275740146637, 0.04535447433590889, -0.18322467803955078, -0.05270809307694435, -0.014244102872908115, 0.03093028999865055, -0.014660698361694813, 0.01945161446928978, 0.10486112534999847, 0.05941476300358772, 0.0038199916016310453, -0.03214489668607712, 0.0883791521191597, 0.00017703020421322435, -0.013874056749045849, -0.02116783708333969, 0.016438279300928116, -0.020714517682790756, 0.027242306619882584, -0.0883953645825386, 0.039983008056879044, -0.08777066320180893, 0.16253571212291718, -0.026616673916578293, 0.01795271225273609, -0.01803608052432537, 0.01510486751794815, -0.04135405272245407, -0.0459599532186985, 0.038306619971990585, 0.040269915014505386, -0.02748258225619793, 0.12608392536640167, -0.12255070358514786, 0.2656562924385071, 0.23153333365917206, -0.24676252901554108, 0.030987726524472237, 0.05082440748810768, -0.029035205021500587, 0.0011753165163099766, -0.03594009950757027, -0.030131207779049873, 0.06072118505835533, 0.02877834625542164, 0.1444566696882248, -0.020607847720384598, -0.030625851824879646, 0.02766864188015461, -0.03944944217801094, -0.007854743860661983, -0.012987053953111172, 0.1553782969713211, -0.1224076896905899, 0.1150563508272171, 0.12474805116653442, -0.17306272685527802, 0.08243760466575623, -0.0462706983089447, -0.05292800813913345, 0.055924903601408005, -0.06441416591405869, 0.0015835707308724523, 0.010405559092760086, -0.14569444954395294, -0.07952921092510223, 0.06186552345752716, -0.0555606372654438, 0.050327349454164505, -0.07673806697130203, -0.003873271867632866, 0.04133608564734459, 0.012367420829832554, 0.023800257593393326, 0.030047839507460594, 0.02440137229859829, 0.02414264716207981, -0.03218723088502884, 0.036073390394449234, 0.03858879953622818, 0.058677006512880325, -0.07215769588947296, 0.19118668138980865, -0.05850012227892876, -0.2645045220851898, -0.23449355363845825, -0.1499338448047638, -0.04403681680560112, 0.06239637732505798, 0.0626073032617569, -0.07654958218336105, -0.026931529864668846, 0.06603135913610458, 0.16042983531951904, -0.06660907715559006, 0.026213601231575012, -0.01779656484723091, 0.027916142717003822, -0.011674294248223305, -0.11765807867050171, -0.022195562720298767, -0.012592166662216187, -0.06822986900806427, 0.04898262768983841, -0.05510227382183075, 0.07460102438926697, 0.17029857635498047, -0.04540790244936943, 0.02849333919584751, 0.009073558263480663, 0.1614820808172226, -0.06610024720430374, -0.03928857669234276, 0.33005163073539734, -0.018648995086550713, -0.0050096819177269936, 0.020368576049804688, 0.02401982992887497, -0.00033106841146945953, 0.005870242603123188, -0.02784726768732071, -0.07958103716373444, -0.21833296120166779, -0.09704166650772095, -0.045471906661987305, -0.021275809034705162, 0.14086923003196716, 0.03940292447805405, 0.14912118017673492, 0.14181914925575256, -0.05000057816505432, 0.009744640439748764, -0.028669213876128197, 0.10064972192049026, 0.2483404278755188, 0.0320095531642437, 0.15249980986118317, -0.058996886014938354, -0.06525474786758423, 0.021307164803147316, -0.03141430765390396, 0.06814437359571457, 0.0131258899345994, 0.016995564103126526, 0.0123975845053792, 0.021353084594011307, 0.09907803684473038, 0.18738462030887604, 0.044949065893888474, -0.035018276423215866, 0.03982667624950409, -0.016631197184324265, -0.10452496260404587, -0.010137037374079227, 0.005957709159702063, -0.02046356163918972, -0.0544988177716732, -0.04393784701824188, 0.027141960337758064, 0.1396958976984024, -0.040748681873083115, -0.3472726345062256, -0.07721024751663208, 0.0003703460097312927, -0.04046925902366638, -0.061051223427057266, 0.02856331132352352, -0.15041382610797882, -0.1234775111079216, 0.03462251275777817, -0.06716755777597427, 0.11523091048002243, -0.11193147301673889, 0.010334562510251999, 0.012558288872241974, 0.02262507937848568, 0.052147578448057175, 0.0724477767944336, -0.27841421961784363, 0.10620254278182983, -0.013066211715340614, 0.057438142597675323, -0.030945025384426117, 0.01904924586415291, 0.040532298386096954, 0.19576753675937653, -0.011210305616259575, -0.011931218206882477, 0.011391074396669865, -0.15584857761859894, -0.02167459763586521, 0.023174496367573738, -0.017899679020047188, -0.04250561073422432, 0.038707226514816284, -0.06917636841535568, -0.031082946807146072, 0.0023761719930917025, -0.0038965807761996984, -0.015774685889482498, -0.11194474995136261, 0.007819734513759613, -0.045939408242702484, 0.11840119957923889, -0.007562050595879555, -0.07955478131771088, -0.06658737361431122, 0.13811393082141876, 0.022131050005555153, -0.1200939193367958, -0.0945863202214241, -0.029332086443901062, 0.0861264020204544, -0.09078091382980347, 0.08225788176059723, -0.08269188553094864, 0.10797368735074997, -0.07425722479820251, -0.19108475744724274, -0.0058015319518744946, -0.15449708700180054, -0.017030581831932068, 0.017551623284816742, 0.10702415555715561, -0.040004655718803406, -0.004049020353704691, 0.029350105673074722, 0.04659685492515564, -0.07581035792827606, -0.06594831496477127, -0.027166780084371567, 0.04477638006210327, 0.026811989024281502, -0.004766638856381178, -0.046914197504520416, 0.058009278029203415, -0.04409220069646835, 0.06022148206830025, 0.20115242898464203, 0.0022116561885923147, -0.03533521667122841, 0.021524330601096153, 0.1987752467393875, -0.05851975083351135, -0.23370984196662903, -0.09723827987909317, -0.009630508720874786, -0.04496308043599129, -0.08511079847812653, -0.18080827593803406, 0.19745266437530518, 0.03145424649119377, -0.03170241415500641, -0.04148704931139946, -0.1878274530172348, -0.06168828532099724, 0.267719566822052, 0.04535006359219551, 0.2193611115217209, -0.2062949240207672, -0.06001090258359909, -0.05239875242114067, -0.15164683759212494, 0.1868363916873932, -0.12545235455036163, 0.05439151078462601, -0.036896031349897385, 0.020621297881007195, 0.023498037829995155, -0.03825949877500534, 0.07677892595529556, -0.035902947187423706, -0.044937923550605774, -0.07003292441368103, 0.05355481803417206, 0.03184172883629799, -0.039177410304546356, 0.08506505191326141, 0.010977853089571, 0.05146056041121483, -0.07874061167240143, -0.07150567322969437, -0.05638188496232033, 0.1446351855993271, 0.03694205731153488, -0.06326215714216232, 0.07436282187700272, -0.09497856348752975, -0.010711227543652058, 0.004925215151160955, 0.11327420920133591, 0.041613202542066574, 0.12990275025367737, 0.17184752225875854, 0.05745600536465645, -0.006229748949408531, -0.1636064499616623, -0.04007580503821373, -0.0626448318362236, 0.11776652932167053, -0.060613274574279785, 0.022434329614043236, 0.052680205553770065, 0.051872577518224716, 0.035178061574697495, 0.054840534925460815, 0.026782305911183357, -0.06533744186162949, 0.06166883558034897, -0.14302553236484528, 0.0029533945489674807, -0.06811351329088211, -0.026450807228684425, 0.020080437883734703, 0.09743563830852509, 0.14990010857582092, -0.058204181492328644, -0.02301039732992649, 0.026730293408036232, -0.06573311984539032, -0.052031561732292175, 0.12331276386976242, 0.05929328873753548, 0.08480270951986313, -0.13230884075164795, 0.041350334882736206, 0.023560693487524986, -0.1412842869758606, -0.02014794386923313, 0.09011349827051163, -0.10503116250038147, -0.11824729293584824, -0.05565419793128967, 0.022689213976264, -0.24039429426193237, -0.08297286182641983, -0.06435106694698334, -0.08259981870651245, 0.0327620804309845, 0.2176225781440735, 0.11985909193754196, 0.015628710389137268, -0.0693126916885376, 0.04410577937960625, -0.12368985265493393, 0.026167109608650208, 0.037127964198589325, 0.08215128630399704, -0.1650434136390686, 0.09427465498447418, 0.036968085914850235, 0.1292790174484253, -0.1038443073630333, -0.03724386915564537, -0.11412860453128815, 0.027631988748908043, -0.13227520883083344, -0.047244951128959656, -0.10184790939092636, -0.013244484551250935, -0.0003032315580639988, -0.06675763428211212, -0.050737760961055756, 0.008746818639338017, -0.07023985683917999, 0.006750944536179304, 0.0021224350202828646, 0.0022188860457390547, -0.1111060306429863, 0.011053390800952911, 0.057541873306035995, -0.05828336626291275, 0.010232911445200443, 0.13505756855010986, -0.039632804691791534, 0.09583493322134018, 0.039078544825315475, -0.10145308822393417, 0.06334823369979858, 0.042299602180719376, 0.01889140158891678, -0.037668514996767044, 0.09198683500289917, 0.06722612679004669, 0.029536211863160133, 0.0060177156701684, 0.04657958447933197, -0.10223402827978134, 0.04373607784509659, 0.010719021782279015, -0.05395786836743355, -0.03760143369436264, -0.019906621426343918, 0.07849522680044174, 0.0632815733551979, 0.1585097759962082, -0.07277626544237137, 0.06323568522930145, -0.09423828125, 0.05045030266046524, -0.060818083584308624, -0.13955192267894745, -0.03341660276055336, -0.07617151737213135, 0.0017660369630903006, 0.03280187025666237, 0.22794939577579498, 0.05762339383363724, 0.14247755706310272, 0.054060667753219604, 0.11442122608423233, 0.03976957127451897, 0.0013395712012425065, 0.28435251116752625, 0.11443108320236206, 0.01575029268860817, -0.021614527329802513, 0.10428906977176666, 0.028158174827694893, -0.06532125920057297, 0.015014546923339367, 0.11979342997074127, -0.0805385410785675, 0.06839337199926376, 0.05540688335895538, 0.019010629504919052, -0.09583955258131027, -0.2483561635017395, -0.029811277985572815, 0.04414952173829079, -0.013506430201232433, 0.14507710933685303, 0.13155809044837952, -0.01927400752902031, -0.012600063346326351, 0.05916227772831917, -0.0343463309109211, -0.15388476848602295, 0.0035581416450440884, -0.06524430215358734, -0.19584132730960846, -0.013086558319628239, -0.06265278160572052, -0.05908820405602455, 0.0013529595453292131, 0.030300291255116463, -0.059371259063482285, 0.09309913218021393, 0.0886770710349083, -0.09025916457176208, 0.0025011319667100906, -0.021382831037044525, 0.020822813734412193, -0.013523434288799763, -0.021342452615499496, -0.030398927628993988, 0.011411014944314957, 0.009859318844974041, 0.009044830687344074, -0.029566984623670578, 0.12598669528961182, -0.07352199405431747, -0.09133797138929367, -0.007269751746207476, 0.06167706102132797, -0.03804430738091469, 0.12454300373792648, -0.018618730828166008, -0.02284812554717064, -0.015299208462238312, 0.2147369086742401, -0.05634567514061928, -0.15765511989593506, -0.09916789084672928, 0.352354496717453, 0.0460105799138546, 0.034389011561870575, 0.014825873076915741, -0.01865905150771141, -0.11510691046714783, 0.3418854773044586, 0.17324043810367584, 0.07259722799062729, 0.02073160745203495, 0.07293989509344101, 0.0058991750702261925, 0.03553306311368942, 0.07980269193649292, 0.13086184859275818, 0.21336545050144196, -0.05351453647017479, -0.07131271064281464, -0.05588983744382858, -0.03289613500237465, -0.08320680260658264, -0.02132556028664112, 0.02708711102604866, -0.06125326454639435, 0.001285756123252213, 0.1051386147737503, -0.187022402882576, 0.016523996368050575, -0.12232333421707153, -0.07505526393651962, -0.03379075974225998, -0.07654917985200882, 0.03504309430718422, -0.0036727951373904943, 0.05217576026916504, 0.008018843829631805, -0.0533861480653286, 0.068658247590065, 0.013593874871730804, -0.17381170392036438, -0.024151738733053207, 0.12398763746023178, 0.009351490996778011, 0.052896611392498016, -0.06353679299354553, 0.041563309729099274, 0.07971470057964325, 0.05128691717982292, -0.04449325054883957, 0.010762677527964115, -0.021161101758480072, -0.17382781207561493, 0.06966635584831238, 0.07072881609201431, -0.02288288064301014, -0.07737409323453903, -0.02383749559521675, -0.26349353790283203, -0.03251701965928078, -0.13623195886611938, 0.0027612901758402586, -0.04816421866416931, 0.030387351289391518, -0.03164094686508179, 0.10233865678310394, 0.13966183364391327, 0.000019038194295717403, -0.01219178456813097, -0.09427793323993683, 0.05444716662168503, 0.07173242419958115, -0.079915851354599, -0.07096809893846512, -0.19548405706882477, -0.08966261148452759, -0.003072933293879032, -0.007424698211252689, -0.16502925753593445, -0.006747741252183914, 0.003068569814786315, -0.02102913334965706, -0.08563804626464844, 0.030682742595672607, 0.09159562736749649, 0.017502987757325172, -0.0014853933826088905, -0.026865413412451744, 0.00042525207391008735, 0.06284230947494507, -0.23236972093582153, -0.08174329996109009 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
automatic-speech-recognition
Imran1/iraqii
[ "transformers", "safetensors", "whisper", "automatic-speech-recognition", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
2024-02-10T15:30:51+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #whisper #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #whisper #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 45, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #whisper #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.05983767658472061, 0.15663617849349976, -0.00414510490372777, 0.012621625326573849, 0.10675175487995148, 0.00396517850458622, 0.07058298587799072, 0.10818448662757874, -0.014333043247461319, 0.1301925629377365, 0.031459614634513855, 0.10620059072971344, 0.11486424505710602, 0.17755427956581116, -0.00021593451674561948, -0.21627318859100342, 0.06542544066905975, -0.11467250436544418, 0.023902224376797676, 0.1205042228102684, 0.14280648529529572, -0.10782013833522797, 0.0710505023598671, -0.02651231922209263, -0.014152529649436474, -0.030523719266057014, -0.05870387330651283, -0.06662651896476746, 0.06516408175230026, 0.0716853216290474, 0.05976768955588341, 0.02008269540965557, 0.07725182175636292, -0.2948664724826813, 0.018899710848927498, 0.0727730244398117, 0.011833904311060905, 0.06048334762454033, 0.07948420196771622, -0.06289119273424149, 0.12036014348268509, -0.044804252684116364, 0.1532549113035202, 0.07767832279205322, -0.09226784855127335, -0.19217613339424133, -0.0771055743098259, 0.06758320331573486, 0.1468338817358017, 0.056199874728918076, -0.03856382891535759, 0.15159031748771667, -0.09224481880664825, 0.0102085517719388, 0.06493527442216873, -0.07805083692073822, -0.04958232864737511, 0.027303149923682213, 0.08463363349437714, 0.08637925982475281, -0.1273571401834488, -0.012682586908340454, 0.03438213840126991, 0.02163512259721756, 0.09837246686220169, 0.025364719331264496, 0.11626957356929779, 0.027283066883683205, -0.13964000344276428, -0.055175989866256714, 0.12345059961080551, 0.033505070954561234, -0.05288216099143028, -0.23939087986946106, -0.010561608709394932, -0.009556320495903492, -0.03001241944730282, -0.04216838628053665, 0.03810601681470871, -0.029798293486237526, 0.07650589942932129, 0.01746492274105549, -0.07078345119953156, -0.04342244938015938, 0.06982958316802979, 0.07824850082397461, 0.022348513826727867, -0.02065650187432766, 0.028734240680933, 0.10911912471055984, 0.08262593299150467, -0.12154309451580048, -0.06694398820400238, -0.06854734569787979, -0.09466245025396347, -0.0454239584505558, 0.03469004109501839, 0.06703099608421326, 0.057105712592601776, 0.19864854216575623, 0.011600262485444546, 0.05358051881194115, 0.022981496527791023, 0.01298176683485508, 0.07163717597723007, 0.07945776730775833, -0.051690056920051575, -0.1315721571445465, -0.04847193509340286, 0.11824512481689453, 0.008524151518940926, -0.033710937947034836, -0.02968421019613743, 0.0653507187962532, 0.05568600073456764, 0.11161840707063675, 0.07554161548614502, 0.01568971388041973, -0.07114148139953613, -0.043046265840530396, 0.19346864521503448, -0.15610936284065247, 0.021089470013976097, 0.019353056326508522, -0.05417651683092117, -0.022803083062171936, 0.007743596564978361, 0.017318524420261383, -0.02697303518652916, 0.1045108512043953, -0.07085666805505753, -0.032245416194200516, -0.1046156957745552, -0.055557940155267715, 0.03224421665072441, 0.009115081280469894, -0.030819423496723175, -0.042374368757009506, -0.09924564510583878, -0.0756484866142273, 0.06214139610528946, -0.07012778520584106, -0.06952599436044693, -0.028100011870265007, -0.04856603220105171, 0.012879165820777416, 0.0010717154946178198, 0.12350035458803177, -0.03162076696753502, 0.043779097497463226, -0.04884343594312668, 0.06864890456199646, 0.13179735839366913, 0.032575443387031555, -0.07970008254051208, 0.058469612151384354, -0.22937731444835663, 0.11186469346284866, -0.09973006695508957, 0.03430512547492981, -0.15810096263885498, -0.02635045349597931, 0.024752190336585045, 0.033622484654188156, -0.017231743782758713, 0.13669319450855255, -0.2039388120174408, -0.036121536046266556, 0.1721590757369995, -0.1349588930606842, -0.08518610149621964, 0.06643460690975189, -0.055845119059085846, 0.11782421916723251, 0.049206800758838654, -0.014434589073061943, 0.04594586789608002, -0.13173595070838928, -0.025916490703821182, -0.053098164498806, -0.007177549879997969, 0.15609249472618103, 0.06614800542593002, -0.06571528315544128, 0.03145577386021614, 0.02247771993279457, -0.018577884882688522, -0.045781973749399185, -0.03384651243686676, -0.09418359398841858, 0.007437155116349459, -0.07286001741886139, 0.00992972869426012, -0.017532840371131897, -0.08721724897623062, -0.039823103696107864, -0.16453123092651367, -0.00716154370456934, 0.09300678223371506, 0.010935397818684578, -0.02714768424630165, -0.09726624190807343, 0.006592306774109602, 0.01717078872025013, -0.01454078033566475, -0.15828220546245575, -0.0459267795085907, 0.03719138726592064, -0.1820053607225418, 0.03403490409255028, -0.05244239792227745, 0.035954125225543976, 0.03684226796030998, -0.03816571831703186, -0.013848266564309597, 0.020031210035085678, 0.018333489075303078, -0.017020072788000107, -0.2371053695678711, -0.014824622310698032, -0.04800339788198471, 0.16693253815174103, -0.23147691786289215, 0.03312116861343384, 0.07037223875522614, 0.12888941168785095, 0.003875810420140624, -0.0490296445786953, 0.030063113197684288, -0.05199332535266876, -0.044617995619773865, -0.05644122138619423, -0.006168664898723364, -0.030205117538571358, -0.04949198290705681, 0.050275903195142746, -0.19857677817344666, -0.041567981243133545, 0.11094366759061813, 0.06673718988895416, -0.1588216871023178, -0.0695650652050972, -0.03473977744579315, -0.06271405518054962, -0.09103205800056458, -0.05391426756978035, 0.10852089524269104, 0.04763965308666229, 0.048611950129270554, -0.07248158007860184, -0.04900932312011719, 0.007940629497170448, -0.00704985111951828, -0.03555170074105263, 0.08515505492687225, 0.08571629226207733, -0.11543579399585724, 0.09118600934743881, 0.06718818843364716, 0.06912244111299515, 0.0983632430434227, -0.0017782750073820353, -0.09694159775972366, -0.014548503793776035, 0.018360106274485588, 0.01051856018602848, 0.12805555760860443, -0.07398705929517746, 0.03667636960744858, 0.05262641981244087, -0.035613641142845154, 0.01095122192054987, -0.101106658577919, 0.029197964817285538, 0.0282101072371006, -0.003792217466980219, 0.028733761981129646, -0.04522410035133362, 0.020432880148291588, 0.1023864597082138, 0.03395526856184006, 0.027725959196686745, 0.010809014551341534, -0.04075441509485245, -0.11779133975505829, 0.1720944494009018, -0.09817105531692505, -0.25773105025291443, -0.12466797232627869, -0.001978461164981127, 0.045932475477457047, -0.018764600157737732, 0.01608397625386715, -0.053159136325120926, -0.11253257840871811, -0.10541603714227676, 0.019763922318816185, 0.058765511959791183, -0.08840499073266983, -0.052470505237579346, 0.04951007664203644, 0.036848895251750946, -0.12439411878585815, 0.021039357408881187, 0.04023430123925209, -0.059992119669914246, 0.0014880987582728267, 0.07059671729803085, 0.08472984284162521, 0.18226684629917145, 0.022740190848708153, -0.01784367859363556, 0.017296429723501205, 0.23125670850276947, -0.1456713229417801, 0.09739834815263748, 0.1370985060930252, -0.06344101577997208, 0.08623462915420532, 0.21197044849395752, 0.036558255553245544, -0.08882707357406616, 0.037767693400382996, 0.03336544707417488, -0.036437466740608215, -0.2318716198205948, -0.08410470932722092, 0.001480261329561472, -0.08248372375965118, 0.0952354297041893, 0.09051923453807831, 0.11156398802995682, 0.04929385334253311, -0.10106591880321503, -0.07701091468334198, 0.04251527413725853, 0.11516540497541428, -0.006902680266648531, 0.004321529995650053, 0.09879171848297119, -0.029613742604851723, 0.010339556261897087, 0.09523830562829971, 0.0004232692008372396, 0.18618540465831757, 0.04265686497092247, 0.12916190922260284, 0.08458086103200912, 0.05236417427659035, 0.02661769837141037, 0.01322705764323473, 0.031609587371349335, 0.02576516941189766, -0.02334577962756157, -0.09271565079689026, -0.012906024232506752, 0.1415313482284546, 0.04929639771580696, 0.030407944694161415, 0.020662572234869003, -0.03531459718942642, 0.07301895320415497, 0.16116659343242645, 0.011933310888707638, -0.21851851046085358, -0.05515235662460327, 0.07743874937295914, -0.08626089245080948, -0.11299191415309906, -0.0025294655933976173, 0.021754881367087364, -0.17833879590034485, 0.05397404730319977, -0.016486117616295815, 0.10160378366708755, -0.11242987960577011, -0.02206907607614994, 0.04055493697524071, 0.07460751384496689, -0.03305850550532341, 0.07621917128562927, -0.20276865363121033, 0.1373196691274643, 0.008098544552922249, 0.06249339506030083, -0.11230216175317764, 0.08414414525032043, 0.019059745594859123, -0.0036223498173058033, 0.1621086448431015, -0.009664713405072689, -0.09406581521034241, -0.060111574828624725, -0.07602227479219437, -0.012445085681974888, 0.09843466430902481, -0.0939253643155098, 0.08608877658843994, -0.01022840291261673, -0.03214890882372856, -0.007143673487007618, -0.11786875873804092, -0.1394684612751007, -0.183831125497818, 0.05997816100716591, -0.10696699470281601, 0.03344186022877693, -0.10895431786775589, -0.060553617775440216, -0.03646453842520714, 0.19020794332027435, -0.18181639909744263, -0.08386372029781342, -0.14476649463176727, -0.07653295993804932, 0.1361350119113922, -0.04076695069670677, 0.07850751280784607, -0.00008746175444684923, 0.20719517767429352, 0.001825421117246151, -0.00039511307841166854, 0.08349475264549255, -0.09573810547590256, -0.20032998919487, -0.0880952924489975, 0.13964824378490448, 0.12494690716266632, 0.04542626440525055, -0.006928097922354937, 0.027518225833773613, -0.011671899817883968, -0.11464269459247589, 0.02507087029516697, 0.1405206173658371, 0.06840235739946365, 0.04314489662647247, -0.016979211941361427, -0.15606153011322021, -0.10666806995868683, -0.05322869494557381, 0.021586019545793533, 0.17797614634037018, -0.07007403671741486, 0.1621050238609314, 0.16129834949970245, -0.05420130863785744, -0.2030099630355835, 0.02282964438199997, 0.04042449966073036, -0.013990761712193489, 0.03615177795290947, -0.19683793187141418, 0.07753707468509674, 0.016794858500361443, -0.060990821570158005, 0.13549083471298218, -0.1619698405265808, -0.1508903205394745, 0.09218499809503555, 0.06408262252807617, -0.2138945758342743, -0.13302136957645416, -0.10209991782903671, -0.05448025092482567, -0.10983701795339584, 0.08582660555839539, 0.01998555287718773, 0.0000906725981622003, 0.04219266399741173, 0.03161109238862991, 0.021054213866591454, -0.0520465187728405, 0.20073460042476654, 0.0012120193568989635, 0.03459459915757179, -0.08232162147760391, -0.08637090027332306, 0.026973288506269455, -0.05251563340425491, 0.0672052875161171, -0.016655180603265762, 0.0002542635484132916, -0.09922616183757782, -0.06439188867807388, -0.06020424887537956, 0.03343502804636955, -0.08179902285337448, -0.09706422686576843, -0.058388181030750275, 0.10227678716182709, 0.08968468755483627, -0.03377925977110863, -0.06091363728046417, -0.10292473435401917, 0.06651771068572998, 0.22872710227966309, 0.1885143369436264, 0.06312023848295212, -0.07107747346162796, 0.0009368667961098254, -0.023646708577871323, 0.050360288470983505, -0.1945972442626953, 0.046965986490249634, 0.042262639850378036, 0.028454279527068138, 0.12927067279815674, -0.024874795228242874, -0.16607771813869476, -0.04733136296272278, 0.06063033267855644, -0.059542834758758545, -0.18076083064079285, -0.000619421829469502, 0.09315520524978638, -0.15953904390335083, -0.06748805940151215, 0.023891208693385124, -0.020897341892123222, -0.027535755187273026, 0.004573860205709934, 0.0820559412240982, 0.02817925252020359, 0.11291294544935226, 0.06535529345273972, 0.10744494199752808, -0.10965088754892349, 0.08151662349700928, 0.09152320772409439, -0.10730767250061035, 0.02777967043220997, 0.07435369491577148, -0.05882004648447037, -0.03269755467772484, 0.0057791233994066715, 0.07514561712741852, 0.02294853888452053, -0.07087770849466324, -0.0009696646011434495, -0.1182747483253479, 0.06833867728710175, 0.13341592252254486, 0.033248964697122574, -0.0019442925695329905, 0.044254120439291, 0.02532937377691269, -0.08849740773439407, 0.11402047425508499, 0.03831348940730095, 0.031180279329419136, -0.04628003388643265, -0.005872894544154406, 0.04073992744088173, -0.011434492655098438, -0.01770744100213051, -0.03857431188225746, -0.061015255749225616, -0.009887747466564178, -0.1567201316356659, 0.02684243768453598, -0.0771009624004364, 0.00816130917519331, 0.022786233574151993, -0.03996667265892029, -0.005420312751084566, 0.006734060123562813, -0.08264576643705368, -0.03730582818388939, -0.0037628922145813704, 0.1070059984922409, -0.15296638011932373, 0.00852613802999258, 0.09225248545408249, -0.12423861026763916, 0.07808402180671692, -0.0011087276507169008, -0.013306759297847748, 0.02074836567044258, -0.1374569684267044, 0.051461800932884216, -0.006391053553670645, 0.011301612481474876, 0.028202330693602562, -0.19194763898849487, 0.0008063786081038415, -0.04062483087182045, -0.05044460669159889, -0.012731820344924927, -0.05135709419846535, -0.11374296247959137, 0.10732509195804596, 0.023315785452723503, -0.08887150883674622, -0.01889934204518795, 0.045546844601631165, 0.10550197213888168, -0.05122669041156769, 0.13676951825618744, -0.01927841641008854, 0.0586048886179924, -0.1769271343946457, -0.014012092724442482, -0.018402719870209694, 0.013554446399211884, -0.017449822276830673, -0.00605781190097332, 0.0551704466342926, -0.012471658177673817, 0.23972837626934052, -0.027916517108678818, 0.03500373288989067, 0.06697984784841537, 0.016924316063523293, -0.018179070204496384, 0.08486920595169067, 0.05455834046006203, 0.026243781670928, 0.01494054775685072, 0.017568159848451614, -0.051871586591005325, -0.021555433049798012, -0.1424977034330368, 0.07956096529960632, 0.16729016602039337, 0.09009124338626862, -0.008234765380620956, 0.06473081558942795, -0.11607895791530609, -0.07983584702014923, 0.10896016657352448, -0.03711748123168945, -0.0032444922253489494, -0.05700715631246567, 0.1502007693052292, 0.1525147259235382, -0.16814833879470825, 0.06879524886608124, -0.06271831691265106, -0.05224054306745529, -0.11435537785291672, -0.16904489696025848, -0.06866718828678131, -0.035694681107997894, -0.002330650808289647, -0.05624498426914215, 0.07767387479543686, 0.10255347937345505, 0.007528870366513729, 0.0038026864640414715, 0.08233556896448135, -0.037537459284067154, -0.006316144950687885, 0.04542352631688118, 0.049430496990680695, 0.015805410221219063, -0.059124622493982315, 0.010986202396452427, 0.004953318741172552, 0.04692067950963974, 0.05509426072239876, 0.034005217254161835, -0.028324270620942116, 0.012686561793088913, -0.018243486061692238, -0.10028578341007233, 0.035927701741456985, -0.033664118498563766, -0.05780354142189026, 0.13973994553089142, 0.0218597874045372, 0.007779987063258886, -0.02196359448134899, 0.22996114194393158, -0.07252145558595657, -0.08971016108989716, -0.1408918797969818, 0.13730354607105255, -0.046912964433431625, 0.05402535945177078, 0.04905577376484871, -0.10465127229690552, 0.0241316556930542, 0.14292258024215698, 0.13702698051929474, -0.027644719928503036, 0.010874779894948006, 0.015687033534049988, 0.00620539765805006, -0.031101418659090996, 0.04872303828597069, 0.04169761762022972, 0.13120494782924652, -0.06359384953975677, 0.0914405807852745, -0.010274309664964676, -0.08765450119972229, -0.0231675673276186, 0.1299583613872528, 0.005232672207057476, 0.02307419292628765, -0.08125553280115128, 0.11583263427019119, -0.0691702738404274, -0.24996554851531982, 0.04865904897451401, -0.05924736708402634, -0.15156961977481842, -0.017320360988378525, 0.02757420763373375, 0.005632835440337658, 0.02303774654865265, 0.06296881288290024, -0.06651590019464493, 0.1557060331106186, 0.035915885120630264, -0.07977382838726044, -0.06385304778814316, 0.08052598685026169, -0.08511006832122803, 0.29178112745285034, 0.010383724234998226, 0.05882499739527702, 0.0948280319571495, -0.028215935453772545, -0.131154403090477, 0.05278646945953369, 0.0955355316400528, -0.07669185847043991, 0.070269875228405, 0.19858962297439575, 0.0003398389380890876, 0.11546503752470016, 0.07913552224636078, -0.09058261662721634, 0.05968843400478363, -0.07367776334285736, -0.09094593673944473, -0.0922231450676918, 0.08578167855739594, -0.06759190559387207, 0.15170368552207947, 0.12874077260494232, -0.043129127472639084, -0.001158626051619649, -0.030579449608922005, 0.051351167261600494, -0.0008969766786321998, 0.12188339978456497, 0.015837527811527252, -0.19386562705039978, 0.031386423856019974, -0.015537483617663383, 0.099497489631176, -0.23898114264011383, -0.07769263535737991, 0.03750690072774887, -0.014490727335214615, -0.048680152744054794, 0.11743341386318207, 0.05373985692858696, 0.045937854796648026, -0.05465031415224075, -0.060885775834321976, 0.006574091035872698, 0.1611197590827942, -0.11137263476848602, 0.004140520468354225 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # opt-1.3b-lora-3.15M-squad-model1 This model is a fine-tuned version of [facebook/opt-1.3b](https://huggingface.co/facebook/opt-1.3b) on the squad dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 48 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.1+cu121 - Datasets 2.15.0 - Tokenizers 0.15.0
{"license": "other", "tags": ["generated_from_trainer"], "datasets": ["varun-v-rao/squad"], "base_model": "facebook/opt-1.3b", "model-index": [{"name": "opt-1.3b-lora-3.15M-squad-model1", "results": []}]}
question-answering
varun-v-rao/opt-1.3b-lora-3.15M-squad-model1
[ "transformers", "tensorboard", "safetensors", "opt", "question-answering", "generated_from_trainer", "dataset:varun-v-rao/squad", "base_model:facebook/opt-1.3b", "license:other", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T15:32:31+00:00
[]
[]
TAGS #transformers #tensorboard #safetensors #opt #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-facebook/opt-1.3b #license-other #endpoints_compatible #text-generation-inference #region-us
# opt-1.3b-lora-3.15M-squad-model1 This model is a fine-tuned version of facebook/opt-1.3b on the squad dataset. ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 48 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 3 ### Training results ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.1+cu121 - Datasets 2.15.0 - Tokenizers 0.15.0
[ "# opt-1.3b-lora-3.15M-squad-model1\n\nThis model is a fine-tuned version of facebook/opt-1.3b on the squad dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 48\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3", "### Training results", "### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0" ]
[ "TAGS\n#transformers #tensorboard #safetensors #opt #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-facebook/opt-1.3b #license-other #endpoints_compatible #text-generation-inference #region-us \n", "# opt-1.3b-lora-3.15M-squad-model1\n\nThis model is a fine-tuned version of facebook/opt-1.3b on the squad dataset.", "## Model description\n\nMore information needed", "## Intended uses & limitations\n\nMore information needed", "## Training and evaluation data\n\nMore information needed", "## Training procedure", "### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 48\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3", "### Training results", "### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0" ]
[ 81, 42, 6, 12, 8, 3, 90, 4, 33 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #opt #question-answering #generated_from_trainer #dataset-varun-v-rao/squad #base_model-facebook/opt-1.3b #license-other #endpoints_compatible #text-generation-inference #region-us \n# opt-1.3b-lora-3.15M-squad-model1\n\nThis model is a fine-tuned version of facebook/opt-1.3b on the squad dataset.## Model description\n\nMore information needed## Intended uses & limitations\n\nMore information needed## Training and evaluation data\n\nMore information needed## Training procedure### Training hyperparameters\n\nThe following hyperparameters were used during training:\n- learning_rate: 2e-05\n- train_batch_size: 16\n- eval_batch_size: 16\n- seed: 48\n- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n- lr_scheduler_type: linear\n- num_epochs: 3### Training results### Framework versions\n\n- Transformers 4.35.2\n- Pytorch 2.1.1+cu121\n- Datasets 2.15.0\n- Tokenizers 0.15.0" ]
[ -0.10573938488960266, 0.16921117901802063, -0.003404607530683279, 0.07416188716888428, 0.11939872056245804, 0.0025775046087801456, 0.10747434198856354, 0.13777396082878113, -0.05493525043129921, 0.09048650413751602, 0.07232332229614258, 0.017564820125699043, 0.0673399418592453, 0.17197783291339874, -0.04640962556004524, -0.17237578332424164, 0.017731154337525368, -0.025736546143889427, -0.0673520639538765, 0.09495016187429428, 0.10040249675512314, -0.09350312501192093, 0.08356814086437225, -0.0011957132956013083, -0.11811710149049759, 0.04422701895236969, -0.016452917829155922, -0.05858674645423889, 0.08297020196914673, 0.03490493446588516, 0.07141769677400589, 0.02961081452667713, 0.10106556117534637, -0.2510935068130493, 0.0063368226401507854, 0.07910013943910599, -0.003908683080226183, 0.07295318692922592, 0.03748662397265434, -0.0028617053758352995, 0.07086451351642609, -0.16785798966884613, 0.10193116962909698, 0.029255704954266548, -0.09774502366781235, -0.18328525125980377, -0.09830330312252045, 0.10377390682697296, 0.10415564477443695, 0.09960629045963287, -0.016710219904780388, 0.141816645860672, -0.05953352153301239, 0.06753182411193848, 0.2027818113565445, -0.255279004573822, -0.056237511336803436, 0.01698305457830429, 0.049434006214141846, 0.06745199859142303, -0.09820560365915298, 0.010425985790789127, 0.06058134138584137, 0.0226677767932415, 0.07969579845666885, -0.005820257589221001, -0.024587564170360565, -0.015707964077591896, -0.12673869729042053, -0.04259698837995529, 0.20217575132846832, 0.0768759474158287, -0.05031796917319298, -0.11669614166021347, -0.040614791214466095, -0.04046717658638954, -0.004211118910461664, -0.06311798840761185, 0.01969911903142929, -0.04863991215825081, -0.05460761487483978, -0.06907594203948975, -0.08958909660577774, -0.03878488764166832, 0.026582062244415283, 0.040470004081726074, 0.03718024492263794, 0.022220483049750328, -0.036138735711574554, 0.07986991107463837, -0.01215582899749279, -0.13797885179519653, -0.014970067888498306, -0.01014635805040598, -0.09492307156324387, -0.06296911835670471, -0.012848583981394768, -0.0290218573063612, 0.009001770056784153, 0.1546124368906021, -0.06911450624465942, 0.05941494554281235, -0.01802767813205719, -0.0021281742956489325, -0.007436980493366718, 0.11328694224357605, -0.03423342853784561, -0.045778706669807434, 0.009580714628100395, 0.116562120616436, 0.022186120972037315, -0.013139592483639717, -0.07360543310642242, -0.018548335880041122, 0.08788279443979263, 0.07792370766401291, -0.004011679440736771, 0.005723880138248205, -0.048357196152210236, -0.02872232347726822, 0.06973475217819214, -0.12342225760221481, 0.049402300268411636, -0.004926449153572321, -0.059717562049627304, -0.08603060245513916, 0.036607950925827026, -0.0009890428045764565, -0.031413108110427856, 0.033440522849559784, -0.06009070202708244, -0.01892971061170101, -0.05045120418071747, -0.043904561549425125, 0.037360984832048416, -0.05723121017217636, -0.007279776968061924, -0.077657550573349, -0.18924658000469208, -0.036938928067684174, 0.023916011676192284, -0.06669206917285919, -0.054936643689870834, -0.01667422614991665, -0.061906956136226654, 0.0029235524125397205, -0.02574363723397255, 0.09349676221609116, -0.04005393013358116, 0.06166590377688408, 0.028050754219293594, 0.033105671405792236, 0.06326578557491302, 0.03967468813061714, -0.07105479389429092, 0.05173952504992485, -0.12965013086795807, 0.07687932997941971, -0.10528109222650528, 0.016200777143239975, -0.1237223818898201, -0.08811511099338531, 0.012771804817020893, -0.02236318401992321, 0.06802310049533844, 0.1291688084602356, -0.16406002640724182, -0.03636433184146881, 0.1579926759004593, -0.10896255075931549, -0.13677488267421722, 0.1243765652179718, -0.04719959571957588, 0.008647718466818333, 0.06198301538825035, 0.15017788112163544, 0.13247261941432953, -0.18165893852710724, -0.04049432650208473, 0.009042346850037575, 0.0516183041036129, 0.018760155886411667, 0.05979059264063835, -0.007594997528940439, 0.06359053403139114, 0.006396979093551636, -0.07600470632314682, -0.002793892053887248, -0.06713731586933136, -0.09010918438434601, -0.05117121338844299, -0.08501490950584412, 0.04012497514486313, 0.05156579613685608, 0.0011822462547570467, -0.09217324107885361, -0.12319419533014297, 0.06954680383205414, 0.1148524060845375, -0.04175746440887451, 0.0045603374019265175, -0.07838137447834015, 0.07751792669296265, -0.0486307330429554, -0.027183180674910545, -0.1770293116569519, -0.12933936715126038, 0.030555281788110733, -0.06966638565063477, 0.018340392038226128, 0.039134588092565536, 0.0565861314535141, 0.05842648074030876, -0.06364109367132187, -0.03595691919326782, -0.10432250797748566, 0.005009016953408718, -0.08155462145805359, -0.1775810420513153, -0.05051850527524948, -0.042724642902612686, 0.17807288467884064, -0.2329881191253662, 0.03139770030975342, 0.027567340061068535, 0.15916304290294647, 0.031213965266942978, -0.04336407780647278, 0.0036533114034682512, 0.008583166636526585, 0.004286468960344791, -0.08893556147813797, 0.02217133343219757, -0.0011549198534339666, -0.07558669149875641, -0.051758307963609695, -0.13373340666294098, 0.0952826738357544, 0.0853266566991806, 0.07283580303192139, -0.08717991411685944, -0.010188057087361813, -0.06401980668306351, -0.027286192402243614, -0.09103476256132126, -0.02584441564977169, 0.18768949806690216, 0.023962046951055527, 0.10764218866825104, -0.0848035141825676, -0.07426389306783676, 0.012155584059655666, -0.013007039204239845, -0.05170269310474396, 0.08146461844444275, 0.04492440074682236, -0.14627261459827423, 0.11679606139659882, 0.1207759901881218, 0.03150761127471924, 0.13040363788604736, -0.05759071931242943, -0.08992457389831543, -0.043184760957956314, 0.032881081104278564, -0.008052940480411053, 0.14360575377941132, -0.08105795830488205, 0.0011775527382269502, 0.026222972199320793, -0.0007701375288888812, 0.025240246206521988, -0.15022125840187073, -0.018196236342191696, 0.03291922062635422, -0.05514386668801308, 0.002897643717005849, -0.024231107905507088, 0.03634233772754669, 0.08691199868917465, 0.023066123947501183, -0.0024157832376658916, 0.026621747761964798, -0.02311377041041851, -0.09551344811916351, 0.16359351575374603, -0.10140129178762436, -0.22001393139362335, -0.11743500083684921, 0.0771329402923584, -0.050197020173072815, -0.02720225416123867, 0.02318890579044819, -0.1011965423822403, -0.06333697587251663, -0.09946385771036148, -0.001634618965908885, -0.009863091632723808, -0.03233633562922478, 0.066923126578331, 0.03536361828446388, 0.11193203926086426, -0.11397125571966171, 0.005400932393968105, -0.010273575782775879, -0.057399023324251175, -0.02744201198220253, 0.04476843401789665, 0.1005866676568985, 0.06751402467489243, -0.016162393614649773, 0.03164716064929962, -0.035773929208517075, 0.2318885177373886, -0.09402002394199371, 0.01034742034971714, 0.14416269958019257, 0.003426405368372798, 0.060841284692287445, 0.12929563224315643, 0.0023854614701122046, -0.101563960313797, 0.03669515252113342, 0.07634090632200241, -0.01292607095092535, -0.24450117349624634, -0.043412018567323685, -0.023311004042625427, -0.04006198048591614, 0.09352816641330719, 0.053456418216228485, 0.030470605939626694, 0.04148603230714798, -0.04065423086285591, 0.0023524302523583174, 0.006746453233063221, 0.0773238092660904, 0.09496666491031647, 0.02776337042450905, 0.0896085649728775, -0.04568953067064285, -0.030389178544282913, 0.07575239986181259, 0.00002011509786825627, 0.24661657214164734, -0.01757187210023403, 0.09928926080465317, 0.040346916764974594, 0.16151678562164307, -0.0378815233707428, 0.0039795939810574055, 0.018417689949274063, -0.008014696650207043, 0.010522786527872086, -0.05510687455534935, -0.021489063277840614, 0.041298940777778625, -0.002988840453326702, 0.031202495098114014, -0.09630294889211655, 0.06453001499176025, 0.035435814410448074, 0.22492334246635437, 0.06685280799865723, -0.2579050064086914, -0.07773783802986145, 0.03042290173470974, -0.028214208781719208, -0.05041207745671272, 0.0023425258696079254, 0.14490477740764618, -0.11931098252534866, 0.08866684883832932, -0.05644851177930832, 0.088294118642807, -0.020253581926226616, 0.00504361093044281, 0.03343687579035759, 0.0678442195057869, -0.002009526127949357, 0.0984693244099617, -0.1812404990196228, 0.21565605700016022, 0.023353349417448044, 0.1021496132016182, -0.07335229963064194, 0.04246772080659866, -0.0009967462392523885, 0.07515104860067368, 0.1478891372680664, -0.003788637463003397, -0.07518015056848526, -0.13771618902683258, -0.08756476640701294, 0.0206401776522398, 0.09211399406194687, -0.049305107444524765, 0.07921590656042099, -0.033545803278684616, -0.0012065491173416376, 0.03740198165178299, -0.007799975574016571, -0.15295352041721344, -0.14159151911735535, 0.014840762130916119, -0.026954365894198418, -0.04988964647054672, -0.10024046152830124, -0.09605373442173004, -0.042595844715833664, 0.18792729079723358, 0.015162239782512188, -0.05623665824532509, -0.14200560748577118, 0.08458001911640167, 0.11244484782218933, -0.07618442922830582, 0.005639682523906231, 0.03548002243041992, 0.15284712612628937, 0.01138224359601736, -0.08685004711151123, 0.05098925903439522, -0.07074005901813507, -0.17718790471553802, -0.0450381264090538, 0.171421080827713, 0.0515020526945591, 0.040214430540800095, 0.029505649581551552, 0.01172649022191763, 0.023561520501971245, -0.08420448750257492, 0.05347982048988342, 0.016714146360754967, 0.09260201454162598, 0.0491715669631958, -0.03382265567779541, -0.02841183915734291, -0.04998430609703064, -0.013605975545942783, 0.10507276654243469, 0.2421659529209137, -0.08070752024650574, 0.08863881975412369, 0.07748597115278244, -0.07579772919416428, -0.1596689522266388, 0.054935354739427567, 0.07569010555744171, 0.006561641115695238, 0.08150714635848999, -0.14825938642024994, 0.09101393073797226, 0.09023671597242355, -0.025454258546233177, 0.023359868675470352, -0.26864826679229736, -0.12941405177116394, 0.0856606513261795, 0.10385466367006302, -0.004875186365097761, -0.13504140079021454, -0.04586721584200859, -0.028927024453878403, -0.13354270160198212, 0.11180117726325989, -0.1360618770122528, 0.06369790434837341, 0.0047275833785533905, 0.06128101795911789, 0.02633407898247242, -0.03782152011990547, 0.14975641667842865, 0.026800451800227165, 0.09051956981420517, -0.06785814464092255, 0.024567145854234695, 0.1328330636024475, -0.06931739300489426, 0.08158915489912033, -0.039249908179044724, 0.06033394858241081, -0.15262378752231598, -0.027644453570246696, -0.07554811239242554, 0.06358175724744797, -0.0745457336306572, -0.043646641075611115, -0.055451132357120514, 0.07269720733165741, 0.06416210532188416, -0.029713086783885956, 0.06294527649879456, 0.019568584859371185, 0.11542132496833801, 0.12697413563728333, 0.11480729281902313, 0.018705356866121292, -0.10753529518842697, 0.001904923701658845, -0.028908545151352882, 0.04071366786956787, -0.10416015237569809, 0.048613689839839935, 0.11203861236572266, 0.023682747036218643, 0.15497806668281555, -0.0013446954544633627, -0.08965833485126495, 0.0011255970457568765, 0.033667489886283875, -0.11864307522773743, -0.18505750596523285, -0.0216554943472147, -0.033100757747888565, -0.15400323271751404, 0.008715084753930569, 0.11387726664543152, -0.04077628627419472, -0.017824530601501465, -0.028339888900518417, 0.04134353995323181, 0.0000856602709973231, 0.14859987795352936, 0.06353557854890823, 0.07183480262756348, -0.07492226362228394, 0.11915401369333267, 0.0681079775094986, -0.0898938849568367, 0.08197549730539322, 0.05054089054465294, -0.07928026467561722, -0.035974953323602676, 0.05481892451643944, 0.1823657602071762, 0.010796482674777508, -0.056358177214860916, -0.09742865711450577, -0.10015678405761719, 0.05571106821298599, 0.11774909496307373, 0.03412918001413345, -0.03553110361099243, 0.006445165723562241, 0.010889417491853237, -0.1271957904100418, 0.1223888024687767, 0.04794997349381447, 0.040539439767599106, -0.1468127816915512, 0.059044696390628815, 0.004330452997237444, 0.04336729273200035, -0.02132590301334858, 0.031023306772112846, -0.08732689172029495, -0.015071462839841843, -0.1511286497116089, -0.001429336261935532, -0.033080946654081345, 0.008018040098249912, -0.026634959504008293, -0.07667990028858185, -0.0457766018807888, 0.0455063171684742, -0.06329698860645294, -0.05612526088953018, 0.029263170436024666, 0.05051295459270477, -0.18467748165130615, -0.03500417247414589, 0.030311228707432747, -0.09177277982234955, 0.08536237478256226, 0.03844674304127693, 0.03736292943358421, 0.0158520620316267, -0.0657690092921257, 0.0006058908766135573, 0.01813199184834957, 0.041937798261642456, 0.05976080521941185, -0.13341271877288818, 0.006583486218005419, -0.011493120342493057, 0.004426578525453806, 0.03295907378196716, 0.03491872549057007, -0.10585591197013855, -0.014740866608917713, -0.04389839991927147, -0.06123853847384453, -0.04784223437309265, 0.06549763679504395, 0.10469493269920349, 0.014675851911306381, 0.15816012024879456, -0.0868038758635521, 0.050857920199632645, -0.22807304561138153, -0.03992484509944916, 0.010579204186797142, -0.01609620824456215, -0.014898820780217648, -0.03174496069550514, 0.0824822261929512, -0.056459978222846985, 0.11986347287893295, -0.0019283383153378963, 0.13026201725006104, 0.054305579513311386, -0.05628063157200813, 0.004783368669450283, 0.015759408473968506, 0.1096610277891159, 0.026280945166945457, -0.01617252454161644, 0.08988194167613983, -0.05106670781970024, 0.06533796340227127, -0.006003976799547672, 0.16366218030452728, 0.1605709046125412, -0.010279054753482342, 0.039505258202552795, 0.09356147795915604, -0.11776787042617798, -0.11958575248718262, 0.057266611605882645, -0.06234422326087952, 0.08704377710819244, -0.03702959790825844, 0.096556656062603, 0.11302325874567032, -0.18492506444454193, 0.05500142648816109, -0.058902889490127563, -0.10142118483781815, -0.10252882540225983, -0.07659479230642319, -0.08273296058177948, -0.10352500528097153, 0.03521061688661575, -0.12842483818531036, 0.0038403377402573824, 0.06289204210042953, -0.0023318163584917784, -0.018835876137018204, 0.1626160591840744, -0.005674086045473814, 0.008706879802048206, 0.06546497344970703, 0.02049442194402218, 0.0022980060894042253, -0.05088271573185921, -0.018420258536934853, 0.060693539679050446, 0.020979946479201317, 0.06309688091278076, -0.027269162237644196, 0.004618991632014513, 0.0338403657078743, -0.00421710591763258, -0.0865577831864357, 0.012087255716323853, 0.007384515833109617, 0.024127263575792313, 0.06652792543172836, 0.05095890909433365, 0.029621191322803497, -0.039202265441417694, 0.23719142377376556, -0.053316500037908554, -0.04892382770776749, -0.147605761885643, 0.1100306436419487, 0.022773010656237602, -0.002289337571710348, 0.071027971804142, -0.12394026666879654, 0.005190625786781311, 0.12987089157104492, 0.13668358325958252, -0.041044168174266815, 0.0055220588110387325, -0.022962119430303574, -0.010293238796293736, -0.03538745269179344, 0.08114202320575714, 0.10014118254184723, 0.02147785760462284, -0.053838349878787994, 0.0011055696522817016, 0.0012169615365564823, -0.041040945798158646, -0.05692721903324127, 0.08753307908773422, 0.017077147960662842, 0.038398757576942444, -0.027215421199798584, 0.08776579797267914, 0.028129849582910538, -0.22018040716648102, 0.050368618220090866, -0.18872331082820892, -0.18198539316654205, -0.01011663954705, 0.08102045953273773, -0.012887095101177692, 0.04422929137945175, 0.0044991509057581425, 0.0013511390425264835, 0.14407552778720856, -0.003985247109085321, -0.06670209020376205, -0.1023881807923317, 0.09044410288333893, -0.11778776347637177, 0.23448516428470612, -0.0005705999792553484, 0.051072828471660614, 0.10470958799123764, -0.02704779803752899, -0.14044733345508575, 0.03236938640475273, 0.08634053170681, -0.019380254670977592, 0.019548336043953896, 0.15832369029521942, -0.03334531560540199, 0.10307174921035767, 0.07898613065481186, -0.10849431902170181, -0.03661090508103371, -0.06198858842253685, -0.004657737910747528, -0.09035803377628326, 0.04381076619029045, -0.07478930801153183, 0.1547105461359024, 0.16840875148773193, -0.048630617558956146, 0.011669113300740719, -0.06408774852752686, 0.03604570031166077, 0.05834973603487015, 0.0996796265244484, 0.010180176235735416, -0.1771901398897171, 0.026375949382781982, 0.044195275753736496, 0.03995135799050331, -0.26455721259117126, -0.09943373501300812, 0.060051098465919495, -0.05031239241361618, -0.05994346737861633, 0.11049293726682663, 0.06407224386930466, 0.04057754948735237, -0.04134316369891167, -0.08457548916339874, -0.05433619022369385, 0.14432427287101746, -0.13715314865112305, -0.018861405551433563 ]
null
null
mlx
<img src="https://l0d0v1c.github.io/notes/medias/greeks.jpg"/> # An Ancient Greek/French Model for RAG * based on mistralai/Mistral-7B-v0.1 * Trained on [1] (and automated translation by GTP4) and [2] * For RAG [3] * Focused on French / Ancient Greek relationship * may produce unaccurate results in translation * https://www.diogenial.com https://rd-mediation.com * Framework MLX Apple Silicon [4] ## Example ### Exemple dans le corpus ``` La traduction de Φασὶ δὲ καὶ Ἀλέξανδρον εἰπεῖν ὡς εἴπερ Ἀλέξανδρος μὴ ἐγεγόνει, ἐθελῆσαι ἂν Διογένης γενέσθαι est: Diogène dit que si Alexandre avait été moins ambitieux, il aurait bien voulu être Diogène. ``` (imprécise: On assure qu’Alexandre disait que s’il n’était pas Alexandre il voudrait être Diogène. selon [5]) ### Exemple hors corpus ``` La traduction de Γνῶθι σεαυτόν est Connais-toi toi-même. ``` ### Exemple de thème hors corpus ``` La traduction de la phrase: J'aime les machines en grec ancien est εἰμὶ δ' ἀγαπῶ τὰ μηχανήματα. ``` Ce qui signifie plus exactement "Je suis et j'aime les machines" ## Use ``` from mlx_lm import load, generate model, tokenizer = load("RANDMEDIATION/DiogenialRAG") response = generate(model, tokenizer, prompt=input("prompt:"), verbose=True,max_tokens=256,temp=0.8) ``` ## References [1] Giannantoni, Gabriele, éd. Socratis et Socraticorum reliquiae. Elenchos 18. Book V. Naples, Italy: Bibliopolis, 1990. [2] https://tatoeba.org/fr/sentences/show_all_in/grc/none [3] Lewis, Patrick, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Naman Goyal, Heinrich Küttler, et al. « Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks ». In Advances in Neural Information Processing Systems, édité par H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, et H. Lin, 33:9459‑74. Curran Associates, Inc., 2020. https://proceedings.neurips.cc/paper_files/paper/2020/file/6b493230205f780e1bc26945df7481e5-Paper.pdf. [4] Hannun, Awni, Jagrit Digani, Angelos Katharopoulos, et Ronan Collobert. « MLX: Efficient and flexible machine learning on Apple silicon », 2023. https://github.com/ml-explore. [5] https://remacle.org/bloodwolf/philosophes/laerce/6diogene1.htm
{"language": ["fr", "el"], "license": "mit", "library_name": "mlx", "tags": ["mlx", "mistral-7b", "Ancient greek"], "pipeline_tag": "text-generation"}
text-generation
RANDMEDIATION/DiogenialRAG
[ "mlx", "mistral-7b", "Ancient greek", "text-generation", "fr", "el", "doi:10.57967/hf/1758", "license:mit", "region:us" ]
2024-02-10T15:33:14+00:00
[]
[ "fr", "el" ]
TAGS #mlx #mistral-7b #Ancient greek #text-generation #fr #el #doi-10.57967/hf/1758 #license-mit #region-us
<img src="URL # An Ancient Greek/French Model for RAG * based on mistralai/Mistral-7B-v0.1 * Trained on [1] (and automated translation by GTP4) and [2] * For RAG [3] * Focused on French / Ancient Greek relationship * may produce unaccurate results in translation * URL URL * Framework MLX Apple Silicon [4] ## Example ### Exemple dans le corpus (imprécise: On assure qu’Alexandre disait que s’il n’était pas Alexandre il voudrait être Diogène. selon [5]) ### Exemple hors corpus ### Exemple de thème hors corpus Ce qui signifie plus exactement "Je suis et j'aime les machines" ## Use ## References [1] Giannantoni, Gabriele, éd. Socratis et Socraticorum reliquiae. Elenchos 18. Book V. Naples, Italy: Bibliopolis, 1990. [2] URL [3] Lewis, Patrick, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Naman Goyal, Heinrich Küttler, et al. « Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks ». In Advances in Neural Information Processing Systems, édité par H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, et H. Lin, 33:9459‑74. Curran Associates, Inc., 2020. URL [4] Hannun, Awni, Jagrit Digani, Angelos Katharopoulos, et Ronan Collobert. « MLX: Efficient and flexible machine learning on Apple silicon », 2023. URL [5] URL
[ "# An Ancient Greek/French Model for RAG\n\n* based on mistralai/Mistral-7B-v0.1\n* Trained on [1] (and automated translation by GTP4) and [2]\n* For RAG [3]\n* Focused on French / Ancient Greek relationship\n* may produce unaccurate results in translation\n* URL URL\n* Framework MLX Apple Silicon [4]", "## Example", "### Exemple dans le corpus\n\n\n(imprécise: On assure qu’Alexandre disait que s’il n’était pas Alexandre il voudrait être Diogène. selon [5])", "### Exemple hors corpus", "### Exemple de thème hors corpus\n\n\nCe qui signifie plus exactement \"Je suis et j'aime les machines\"", "## Use", "## References\n[1] Giannantoni, Gabriele, éd. Socratis et Socraticorum reliquiae. Elenchos 18. Book V. Naples, Italy: Bibliopolis, 1990.\n\n[2] URL\n\n[3] Lewis, Patrick, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Naman Goyal, Heinrich Küttler, et al. « Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks ». In Advances in Neural Information Processing Systems, édité par H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, et H. Lin, 33:9459‑74. Curran Associates, Inc., 2020. URL\n\n[4] Hannun, Awni, Jagrit Digani, Angelos Katharopoulos, et Ronan Collobert. « MLX: Efficient and flexible machine learning on Apple silicon », 2023. URL\n\n[5] URL" ]
[ "TAGS\n#mlx #mistral-7b #Ancient greek #text-generation #fr #el #doi-10.57967/hf/1758 #license-mit #region-us \n", "# An Ancient Greek/French Model for RAG\n\n* based on mistralai/Mistral-7B-v0.1\n* Trained on [1] (and automated translation by GTP4) and [2]\n* For RAG [3]\n* Focused on French / Ancient Greek relationship\n* may produce unaccurate results in translation\n* URL URL\n* Framework MLX Apple Silicon [4]", "## Example", "### Exemple dans le corpus\n\n\n(imprécise: On assure qu’Alexandre disait que s’il n’était pas Alexandre il voudrait être Diogène. selon [5])", "### Exemple hors corpus", "### Exemple de thème hors corpus\n\n\nCe qui signifie plus exactement \"Je suis et j'aime les machines\"", "## Use", "## References\n[1] Giannantoni, Gabriele, éd. Socratis et Socraticorum reliquiae. Elenchos 18. Book V. Naples, Italy: Bibliopolis, 1990.\n\n[2] URL\n\n[3] Lewis, Patrick, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Naman Goyal, Heinrich Küttler, et al. « Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks ». In Advances in Neural Information Processing Systems, édité par H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, et H. Lin, 33:9459‑74. Curran Associates, Inc., 2020. URL\n\n[4] Hannun, Awni, Jagrit Digani, Angelos Katharopoulos, et Ronan Collobert. « MLX: Efficient and flexible machine learning on Apple silicon », 2023. URL\n\n[5] URL" ]
[ 45, 78, 3, 42, 7, 25, 2, 212 ]
[ "passage: TAGS\n#mlx #mistral-7b #Ancient greek #text-generation #fr #el #doi-10.57967/hf/1758 #license-mit #region-us \n# An Ancient Greek/French Model for RAG\n\n* based on mistralai/Mistral-7B-v0.1\n* Trained on [1] (and automated translation by GTP4) and [2]\n* For RAG [3]\n* Focused on French / Ancient Greek relationship\n* may produce unaccurate results in translation\n* URL URL\n* Framework MLX Apple Silicon [4]## Example### Exemple dans le corpus\n\n\n(imprécise: On assure qu’Alexandre disait que s’il n’était pas Alexandre il voudrait être Diogène. selon [5])### Exemple hors corpus### Exemple de thème hors corpus\n\n\nCe qui signifie plus exactement \"Je suis et j'aime les machines\"## Use## References\n[1] Giannantoni, Gabriele, éd. Socratis et Socraticorum reliquiae. Elenchos 18. Book V. Naples, Italy: Bibliopolis, 1990.\n\n[2] URL\n\n[3] Lewis, Patrick, Ethan Perez, Aleksandra Piktus, Fabio Petroni, Vladimir Karpukhin, Naman Goyal, Heinrich Küttler, et al. « Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks ». In Advances in Neural Information Processing Systems, édité par H. Larochelle, M. Ranzato, R. Hadsell, M. F. Balcan, et H. Lin, 33:9459‑74. Curran Associates, Inc., 2020. URL\n\n[4] Hannun, Awni, Jagrit Digani, Angelos Katharopoulos, et Ronan Collobert. « MLX: Efficient and flexible machine learning on Apple silicon », 2023. URL\n\n[5] URL" ]
[ -0.04718749597668648, 0.16278371214866638, -0.007858059369027615, 0.09407863020896912, 0.05676746740937233, -0.009048638865351677, 0.13643987476825714, 0.08686868101358414, 0.13827848434448242, 0.03979433700442314, 0.038515854626894, 0.13059720396995544, 0.04866480454802513, -0.03851406276226044, 0.0022347148042172194, -0.2236710488796234, 0.0466604083776474, -0.020705433562397957, -0.024699170142412186, 0.047382645308971405, 0.07740005850791931, -0.017806513234972954, 0.09208308905363083, -0.08052267134189606, 0.007636304013431072, -0.015447448007762432, -0.05103336647152901, 0.00047005005762912333, 0.11499479413032532, 0.06381136178970337, 0.02712063118815422, 0.018375521525740623, 0.04852282628417015, -0.094509556889534, 0.018031971529126167, 0.02642463892698288, 0.0019777051638811827, -0.012277502566576004, 0.027784397825598717, -0.06585639715194702, 0.17222408950328827, -0.0021737082861363888, 0.01456748228520155, -0.02467128075659275, -0.14069117605686188, 0.02707265131175518, -0.11649131774902344, -0.0497780404984951, 0.0037361353170126677, 0.07396405190229416, 0.0014513609930872917, 0.054636020213365555, 0.0013093337183818221, 0.05068580433726311, 0.14444135129451752, -0.1472788155078888, -0.07373849302530289, 0.04023551195859909, 0.07826663553714752, 0.16127458214759827, -0.0912306010723114, 0.06453990191221237, 0.027818070724606514, -0.01800776831805706, -0.02187385782599449, -0.10088574886322021, 0.10880385339260101, 0.0008967703906819224, -0.11712582409381866, -0.03418906778097153, 0.20972372591495514, 0.001402075868099928, -0.06531623005867004, 0.04138508811593056, -0.027978956699371338, 0.12788814306259155, 0.012068959884345531, -0.12875531613826752, -0.004535230342298746, -0.027958599850535393, 0.17272354662418365, -0.12016460299491882, -0.039462558925151825, 0.005230490583926439, -0.0704130306839943, 0.15964241325855255, 0.055652327835559845, -0.023051390424370766, 0.08703246712684631, 0.02870754897594452, 0.013798343017697334, -0.019471323117613792, -0.0022880560718476772, -0.005707079544663429, -0.007606071885675192, -0.011379186064004898, -0.009173442609608173, 0.050752826035022736, 0.009111077524721622, 0.13203983008861542, -0.05414281785488129, -0.017336692661046982, -0.0027924918103963137, 0.036486588418483734, 0.020079484209418297, 0.04601006582379341, -0.0763152465224266, -0.18273040652275085, -0.00044743239413946867, 0.04175123944878578, -0.04134630039334297, -0.02384229749441147, -0.06831689178943634, -0.03133365139365196, -0.031312596052885056, 0.041370149701833725, 0.07033803313970566, -0.014736561104655266, -0.02583444118499756, -0.04500178247690201, -0.051214925944805145, -0.09636078029870987, -0.024374715983867645, -0.030217567458748817, -0.1256066858768463, 0.08900178968906403, -0.051638901233673096, 0.036455828696489334, -0.14216348528862, -0.0041412427090108395, -0.03451145067811012, 0.01711609959602356, -0.0318286195397377, -0.043699201196432114, 0.024669593200087547, -0.10603231191635132, 0.030225174501538277, -0.11393576115369797, -0.18013575673103333, -0.08561918139457703, 0.05052390322089195, -0.09521334618330002, 0.020172828808426857, -0.0553300566971302, -0.019418038427829742, -0.037843406200408936, 0.03252705931663513, -0.06697499006986618, -0.050577208399772644, 0.05988549068570137, -0.12442851066589355, 0.01918014883995056, -0.07104853540658951, 0.044688861817121506, -0.07804176956415176, -0.022266538813710213, -0.1074821949005127, 0.04293433576822281, -0.08534575998783112, -0.02744920179247856, -0.1393308937549591, -0.00855114683508873, -0.034007180482149124, 0.03846735134720802, 0.03644329309463501, 0.1304437667131424, -0.1363087296485901, -0.03916902095079422, 0.017466312274336815, -0.03150177001953125, -0.07020708173513412, 0.21537072956562042, 0.010416599921882153, 0.13557378947734833, 0.09861885011196136, 0.1481132209300995, 0.016611358150839806, -0.02910064160823822, -0.03860660269856453, -0.06946583837270737, 0.08617625385522842, 0.07827845215797424, 0.060481421649456024, 0.011212688870728016, 0.07528777420520782, 0.024219723418354988, -0.12957249581813812, -0.08408022671937943, -0.009228689596056938, -0.06561152637004852, 0.023653296753764153, -0.04091702401638031, 0.040951918810606, 0.03455141559243202, 0.03622103109955788, 0.03211484104394913, -0.05798390135169029, -0.012700370512902737, 0.01185540109872818, -0.0018374994397163391, 0.016390331089496613, -0.05298765376210213, 0.18312928080558777, 0.06951210647821426, 0.04627399891614914, -0.08348196744918823, 0.012530440464615822, 0.07512146979570389, -0.030016189441084862, 0.16471131145954132, 0.07343458384275436, 0.037756696343421936, 0.09098184108734131, -0.03820472210645676, 0.03347603231668472, -0.06828904896974564, -0.0076195974834263325, -0.038933608680963516, -0.12003567069768906, 0.10573785752058029, -0.031858671456575394, 0.08251799643039703, -0.08677952736616135, 0.033651817589998245, 0.046389248222112656, 0.12670239806175232, -0.02571003884077072, -0.02506793476641178, -0.08615508675575256, 0.010670505464076996, -0.04480968788266182, -0.0036036279052495956, 0.055428266525268555, -0.05216392129659653, -0.04619184508919716, 0.035352468490600586, -0.07748373597860336, 0.06387148052453995, 0.06279600411653519, -0.03509364277124405, -0.045319173485040665, -0.05807848647236824, -0.019913073629140854, -0.021856844425201416, 0.05014017969369888, -0.06974966078996658, 0.20357194542884827, 0.019378187134861946, 0.02606094442307949, -0.13066929578781128, -0.022459080442786217, 0.005101645831018686, -0.01001293957233429, -0.05395640432834625, 0.13928046822547913, 0.09855608642101288, -0.12960076332092285, 0.1077561229467392, 0.17025619745254517, -0.06309094280004501, 0.18906427919864655, 0.023544592782855034, -0.07648850232362747, -0.07763563096523285, 0.07465467602014542, -0.006773028988391161, 0.05273402854800224, -0.10725048929452896, 0.039340727031230927, -0.028426077216863632, -0.0018919536378234625, 0.02314065583050251, -0.08712746948003769, -0.01969335228204727, -0.06030614674091339, -0.00673376489430666, 0.13102707266807556, -0.0004124372499063611, -0.035213902592659, 0.1277400255203247, 0.005824587773531675, -0.12970410287380219, -0.04173732548952103, -0.013842419721186161, -0.019154822453856468, 0.11425578594207764, -0.07074110954999924, -0.13783180713653564, -0.051232077181339264, 0.007660508621484041, -0.14567390084266663, 0.009680528193712234, -0.0011178809218108654, -0.007117515429854393, -0.005584074649959803, -0.025769270956516266, 0.09876254945993423, 0.0025567554403096437, -0.06449833512306213, -0.06625184416770935, 0.013521676883101463, -0.10064330697059631, -0.03408866748213768, -0.03490366041660309, -0.02296091429889202, -0.13460901379585266, -0.025983547791838646, -0.09019194543361664, 0.0944499522447586, 0.011385440826416016, 0.018994495272636414, -0.0426371730864048, -0.0016808264190331101, 0.22076351940631866, -0.1224750503897667, 0.12402765452861786, 0.06961634010076523, -0.013291135430335999, 0.06978102773427963, 0.13545136153697968, 0.009618079289793968, -0.08217104524374008, -0.01742376573383808, 0.06478582322597504, -0.039575960487127304, -0.2562306225299835, -0.1089424341917038, -0.05775529146194458, 0.11267784982919693, 0.053364869207143784, 0.030830079689621925, 0.04420044273138046, 0.027615079656243324, -0.10886494070291519, 0.0735817477107048, 0.12313283234834671, 0.038432106375694275, 0.12124980241060257, -0.022911934182047844, 0.058451537042856216, -0.002050781389698386, -0.03839297965168953, 0.10264481604099274, 0.17661520838737488, 0.15335401892662048, 0.04891233146190643, 0.1040835753083229, 0.020855525508522987, 0.06707970052957535, -0.06798358261585236, -0.0036116382107138634, 0.06814166903495789, 0.0296185202896595, -0.07750026881694794, -0.043296921998262405, 0.030687740072607994, 0.1338018774986267, 0.03763788938522339, -0.10040968656539917, -0.006925668101757765, 0.05714653804898262, 0.07330988347530365, 0.1267068237066269, -0.037414222955703735, -0.1306481808423996, -0.06350988894701004, 0.07391877472400665, -0.0045156474225223064, -0.057591333985328674, 0.04834479093551636, 0.04079641401767731, -0.1123662069439888, 0.011868748813867569, 0.004073892254382372, 0.06262464076280594, -0.11568643152713776, 0.04258468374609947, -0.10871525853872299, -0.027015816420316696, -0.0008610847289673984, 0.03633721545338631, -0.17056825757026672, 0.24533170461654663, -0.001538073760457337, 0.044187020510435104, -0.030816182494163513, -0.016382092610001564, 0.06813531368970871, 0.0681610256433487, 0.1718817949295044, 0.02996988222002983, -0.09188642352819443, -0.04848364368081093, 0.0048396349884569645, -0.03858683258295059, 0.10202962160110474, -0.11149462312459946, 0.05521552637219429, -0.060869425535202026, 0.03683588281273842, -0.05471789836883545, -0.03273400291800499, -0.08486975729465485, -0.16189423203468323, 0.1043284460902214, -0.11034291982650757, -0.00683612423017621, 0.007387914229184389, -0.08678008615970612, -0.11991295218467712, 0.1586749106645584, -0.16342197358608246, 0.010358194820582867, -0.057800959795713425, 0.01499238982796669, 0.07493104785680771, -0.07763891667127609, -0.02812701277434826, -0.04653908684849739, -0.004854004830121994, -0.09397879242897034, 0.03690788149833679, 0.04057573899626732, -0.043337926268577576, -0.0949256494641304, -0.05619511753320694, 0.059154123067855835, 0.08586107194423676, 0.04698629304766655, 0.04355384409427643, 0.0002924448635894805, 0.012358029372990131, -0.09717931598424911, 0.045573171228170395, 0.016831155866384506, -0.04173019155859947, 0.08244139701128006, -0.1674327254295349, -0.09979187697172165, -0.13351301848888397, -0.0914003849029541, 0.02698003686964512, 0.14784802496433258, 0.018603354692459106, 0.02338387444615364, 0.20270664989948273, -0.13864640891551971, -0.2378474622964859, -0.041147030889987946, -0.013855946250259876, -0.041843146085739136, 0.008120015263557434, -0.1730833500623703, 0.03964272141456604, 0.14097942411899567, 0.03305463865399361, -0.06077045574784279, -0.19970661401748657, -0.07752109318971634, 0.018996044993400574, 0.08748690783977509, 0.11165578663349152, -0.06510451436042786, -0.024394337087869644, -0.020247461274266243, 0.004630610346794128, 0.05912596359848976, 0.05424274504184723, 0.07141757756471634, 0.005622156895697117, 0.009852119721472263, -0.008920146152377129, 0.015139743685722351, 0.1539335995912552, 0.006146065890789032, -0.00984107330441475, -0.06259475648403168, -0.09589926153421402, 0.0513041727244854, 0.02145649679005146, 0.02362763322889805, 0.03368688374757767, -0.03668135404586792, -0.055523548275232315, -0.013916248455643654, -0.029967058449983597, 0.08969991654157639, -0.09303116053342819, -0.050407666712999344, -0.06316913664340973, 0.13412460684776306, 0.03125758469104767, 0.0233570896089077, -0.007923215627670288, -0.0058717988431453705, 0.11322872340679169, 0.09983184933662415, 0.07214764505624771, 0.011500596068799496, 0.06129002943634987, -0.027608396485447884, 0.01894560642540455, 0.015170067548751831, -0.0039052581414580345, 0.0006608563708141446, 0.20327416062355042, -0.030771438032388687, 0.11980445683002472, -0.005101518239825964, -0.12844957411289215, 0.0025335487443953753, 0.009982598014175892, -0.0907595306634903, -0.2519807517528534, -0.039081182330846786, -0.05416202172636986, 0.031598687171936035, 0.015004894696176052, 0.22390145063400269, -0.06817161291837692, -0.024705808609724045, -0.01831388659775257, 0.07880061119794846, 0.012096093967556953, 0.10349776595830917, -0.05507616698741913, -0.011331954039633274, -0.03057669661939144, 0.12722893059253693, 0.10275963693857193, -0.13768470287322998, 0.024739742279052734, 0.09333167970180511, -0.036375269293785095, -0.08663275837898254, -0.10337676107883453, 0.016247935593128204, -0.1704469472169876, 0.005670738406479359, -0.09067177027463913, -0.10127058625221252, 0.04545098915696144, 0.10092659294605255, 0.023898471146821976, 0.007431501056998968, -0.016164187341928482, -0.0009352470515295863, 0.010165787301957607, 0.08175650984048843, 0.08391668647527695, -0.06176057085394859, 0.04870854690670967, -0.05558677017688751, -0.01694522239267826, -0.02921496145427227, 0.0007640520343557, -0.039730727672576904, -0.14547160267829895, 0.00016047345707193017, -0.10031837970018387, 0.0012557819718495011, -0.06054389849305153, 0.013939577154815197, -0.03694776073098183, -0.10927309095859528, -0.05692629888653755, -0.01634703390300274, -0.08711891621351242, 0.018237289041280746, 0.02348034456372261, 0.1322839856147766, -0.07972894608974457, -0.01740219071507454, 0.09083645045757294, -0.02716522105038166, 0.05093785747885704, 0.02201811969280243, -0.006150882225483656, 0.09383077174425125, -0.1439383327960968, 0.13143542408943176, -0.08002807945013046, 0.0738154947757721, 0.006645443849265575, -0.08817656338214874, -0.024736251682043076, 0.02334795519709587, 0.044520482420921326, 0.012858912348747253, 0.015943799167871475, -0.014345228672027588, 0.020082484930753708, 0.019250240176916122, -0.12740999460220337, -0.05252719670534134, 0.09330807626247406, 0.07025279104709625, 0.04343710094690323, 0.05102923884987831, 0.014949709177017212, 0.03255720064043999, -0.06066054478287697, 0.05114511027932167, 0.027605876326560974, -0.1292468011379242, -0.04803633689880371, -0.0778731033205986, 0.0434701144695282, 0.04102400690317154, 0.044593024998903275, -0.01076411735266447, 0.06239863857626915, 0.046171918511390686, 0.015714308246970177, -0.12279549986124039, 0.03911178186535835, -0.09134246408939362, 0.06840024888515472, -0.013511087745428085, -0.09242914617061615, -0.015985781326889992, -0.009391971863806248, 0.10556451976299286, 0.06097454950213432, 0.1521366983652115, 0.16175900399684906, 0.06629150360822678, 0.039854247123003006, 0.001028195838443935, -0.04244428500533104, 0.06006128713488579, 0.09851548820734024, 0.03693505749106407, -0.10741695016622543, 0.07485000789165497, 0.1664808690547943, -0.15790966153144836, 0.07470613718032837, -0.006333828903734684, -0.07138844579458237, -0.06471268087625504, -0.20533247292041779, -0.016185428947210312, 0.031208420172333717, -0.03813978284597397, -0.10861219465732574, 0.07008182257413864, -0.030362362042069435, 0.05052351951599121, -0.07301291078329086, 0.03259701654314995, -0.050575148314237595, -0.09653192013502121, 0.1070413663983345, 0.03475043550133705, 0.05131066218018532, -0.04071616381406784, -0.06524870544672012, -0.0890992283821106, 0.04598581790924072, -0.02199888415634632, 0.03328246995806694, -0.015088193118572235, -0.0240416768938303, -0.08096511662006378, -0.023451736196875572, 0.000269539188593626, -0.013675719499588013, -0.05551576241850853, 0.1222522109746933, 0.03583381697535515, -0.04605376720428467, 0.04991481080651283, 0.23063188791275024, 0.03707248345017433, 0.008563803508877754, -0.1045244038105011, 0.049582745879888535, 0.06781796365976334, 0.05480720102787018, 0.0022540895733982325, -0.01717200130224228, 0.03248647227883339, 0.1535460203886032, 0.1441853940486908, -0.0712348222732544, -0.011339502409100533, 0.02579585835337639, 0.06852733343839645, 0.05027147755026817, 0.0038394031580537558, 0.06185414642095566, 0.207449272274971, -0.04716261103749275, 0.05404302477836609, -0.04557148739695549, 0.0035778656601905823, -0.05165887251496315, 0.06234520673751831, 0.0024635889567434788, 0.007211238145828247, -0.07158968597650528, 0.11240264773368835, -0.1416996717453003, -0.29516908526420593, -0.06355523318052292, -0.16372939944267273, -0.10468202829360962, -0.037168022245168686, -0.012430784292519093, 0.01795262284576893, 0.09055548161268234, 0.03268112242221832, -0.06971775740385056, 0.021591683849692345, 0.04850711673498154, -0.07734102755784988, -0.06721201539039612, 0.05795890837907791, 0.028318360447883606, 0.17970654368400574, -0.030228298157453537, 0.06106423959136009, 0.08206839859485626, 0.008118541911244392, -0.028651142492890358, 0.009970229119062424, 0.09942701458930969, -0.02101958729326725, -0.002555668819695711, 0.06088021397590637, 0.002333062468096614, 0.1216319128870964, 0.12710171937942505, 0.06182447820901871, 0.074095219373703, 0.09762275964021683, -0.10954824835062027, 0.041728079319000244, 0.13606750965118408, -0.16462011635303497, 0.11961700767278671, 0.15309596061706543, 0.018133599311113358, 0.043581899255514145, -0.07070069760084152, 0.016148606315255165, 0.01834562048316002, 0.0677470862865448, 0.01258639246225357, -0.21835307776927948, 0.00222472520545125, -0.06960170716047287, 0.07710618525743484, -0.04039010405540466, -0.11169534176588058, -0.006987258791923523, 0.030427951365709305, -0.08080288767814636, 0.11079848557710648, 0.08147697150707245, -0.0920817106962204, -0.020963454619050026, -0.2264889031648636, -0.042564600706100464, 0.06983938068151474, -0.003299436764791608, -0.020951109007000923 ]
null
null
transformers
# LoftQ Initialization | [Paper](https://arxiv.org/abs/2310.08659) | [Code](https://github.com/yxli2123/LoftQ) | [PEFT Example](https://github.com/huggingface/peft/tree/main/examples/loftq_finetuning) | LoftQ (LoRA-fine-tuning-aware Quantization) provides a quantized backbone Q and LoRA adapters A and B, given a full-precision pre-trained weight W. This model, `Mistral-7B-v0.1-4bit-64rank`, is obtained from [Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1). The backbone is under `LoftQ/Mistral-7B-v0.1-4bit-64rank` and LoRA adapters are under the `subfolder='loftq_init'`. ## Model Info ### Backbone - Stored format: `torch.bfloat16` - Size: ~ 14 GiB - Loaded format: bitsandbytes nf4 - Size loaded on GPU: ~3.5 GiB ### LoRA adapters - rank: 64 - lora_alpha: 16 - target_modules: ["down_proj", "up_proj", "q_proj", "k_proj", "v_proj", "o_proj", "gate_proj"] ## Usage **Training.** Here's an example of loading this model and preparing for the LoRA fine-tuning. ```python import torch from transformers import AutoModelForCausalLM, BitsAndBytesConfig from peft import PeftModel MODEL_ID = "LoftQ/Mistral-7B-v0.1-4bit-64rank" base_model = AutoModelForCausalLM.from_pretrained( MODEL_ID, torch_dtype=torch.bfloat16, # you may change it with different models quantization_config=BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_compute_dtype=torch.bfloat16, # bfloat16 is recommended bnb_4bit_use_double_quant=False, bnb_4bit_quant_type='nf4', ), ) peft_model = PeftModel.from_pretrained( base_model, MODEL_ID, subfolder="loftq_init", is_trainable=True, ) # Do training with peft_model ... ``` **Inference.** Here is an example code for inference after the model has been fine-tuned on [GSM8K](https://huggingface.co/datasets/gsm8k). ```python import torch from transformers import AutoModelForCausalLM, BitsAndBytesConfig from peft import PeftModel MODEL_ID = "LoftQ/Mistral-7B-v0.1-4bit-64rank" base_model = AutoModelForCausalLM.from_pretrained( MODEL_ID, torch_dtype=torch.bfloat16, # you may change it with different models quantization_config=BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_compute_dtype=torch.bfloat16, # bfloat16 is recommended bnb_4bit_use_double_quant=False, bnb_4bit_quant_type='nf4', ), ) peft_model = PeftModel.from_pretrained( base_model, MODEL_ID, subfolder="gsm8k", is_trainable=True, ) # Do inference with peft_model ... ``` See the full code at our [Github Repo]((https://github.com/yxli2123/LoftQ)) ## Citation ```bibtex @article{li2023loftq, title={Loftq: Lora-fine-tuning-aware quantization for large language models}, author={Li, Yixiao and Yu, Yifan and Liang, Chen and He, Pengcheng and Karampatziakis, Nikos and Chen, Weizhu and Zhao, Tuo}, journal={arXiv preprint arXiv:2310.08659}, year={2023} } ```
{"language": ["en"], "license": "mit", "tags": ["quantization ", "lora"], "pipeline_tag": "text-generation"}
text-generation
ChuGyouk/Mistral-7B-v0.1-4bit-64rank
[ "transformers", "safetensors", "mistral", "text-generation", "quantization ", "lora", "en", "arxiv:2310.08659", "license:mit", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T15:34:18+00:00
[ "2310.08659" ]
[ "en" ]
TAGS #transformers #safetensors #mistral #text-generation #quantization #lora #en #arxiv-2310.08659 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# LoftQ Initialization | Paper | Code | PEFT Example | LoftQ (LoRA-fine-tuning-aware Quantization) provides a quantized backbone Q and LoRA adapters A and B, given a full-precision pre-trained weight W. This model, 'Mistral-7B-v0.1-4bit-64rank', is obtained from Mistral-7B-v0.1. The backbone is under 'LoftQ/Mistral-7B-v0.1-4bit-64rank' and LoRA adapters are under the 'subfolder='loftq_init''. ## Model Info ### Backbone - Stored format: 'torch.bfloat16' - Size: ~ 14 GiB - Loaded format: bitsandbytes nf4 - Size loaded on GPU: ~3.5 GiB ### LoRA adapters - rank: 64 - lora_alpha: 16 - target_modules: ["down_proj", "up_proj", "q_proj", "k_proj", "v_proj", "o_proj", "gate_proj"] ## Usage Training. Here's an example of loading this model and preparing for the LoRA fine-tuning. Inference. Here is an example code for inference after the model has been fine-tuned on GSM8K. See the full code at our Github Repo)
[ "# LoftQ Initialization\n\n| Paper | Code | PEFT Example |\n\nLoftQ (LoRA-fine-tuning-aware Quantization) provides a quantized backbone Q and LoRA adapters A and B, given a full-precision pre-trained weight W.\n\nThis model, 'Mistral-7B-v0.1-4bit-64rank', is obtained from Mistral-7B-v0.1. \nThe backbone is under 'LoftQ/Mistral-7B-v0.1-4bit-64rank' and LoRA adapters are under the 'subfolder='loftq_init''.", "## Model Info", "### Backbone\n- Stored format: 'torch.bfloat16'\n- Size: ~ 14 GiB\n- Loaded format: bitsandbytes nf4\n- Size loaded on GPU: ~3.5 GiB", "### LoRA adapters\n- rank: 64\n- lora_alpha: 16\n- target_modules: [\"down_proj\", \"up_proj\", \"q_proj\", \"k_proj\", \"v_proj\", \"o_proj\", \"gate_proj\"]", "## Usage\n\nTraining. Here's an example of loading this model and preparing for the LoRA fine-tuning.\n\n\n\nInference. Here is an example code for inference after the model has been fine-tuned on GSM8K.\n\n\n\nSee the full code at our Github Repo)" ]
[ "TAGS\n#transformers #safetensors #mistral #text-generation #quantization #lora #en #arxiv-2310.08659 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# LoftQ Initialization\n\n| Paper | Code | PEFT Example |\n\nLoftQ (LoRA-fine-tuning-aware Quantization) provides a quantized backbone Q and LoRA adapters A and B, given a full-precision pre-trained weight W.\n\nThis model, 'Mistral-7B-v0.1-4bit-64rank', is obtained from Mistral-7B-v0.1. \nThe backbone is under 'LoftQ/Mistral-7B-v0.1-4bit-64rank' and LoRA adapters are under the 'subfolder='loftq_init''.", "## Model Info", "### Backbone\n- Stored format: 'torch.bfloat16'\n- Size: ~ 14 GiB\n- Loaded format: bitsandbytes nf4\n- Size loaded on GPU: ~3.5 GiB", "### LoRA adapters\n- rank: 64\n- lora_alpha: 16\n- target_modules: [\"down_proj\", \"up_proj\", \"q_proj\", \"k_proj\", \"v_proj\", \"o_proj\", \"gate_proj\"]", "## Usage\n\nTraining. Here's an example of loading this model and preparing for the LoRA fine-tuning.\n\n\n\nInference. Here is an example code for inference after the model has been fine-tuned on GSM8K.\n\n\n\nSee the full code at our Github Repo)" ]
[ 69, 140, 3, 49, 68, 65 ]
[ "passage: TAGS\n#transformers #safetensors #mistral #text-generation #quantization #lora #en #arxiv-2310.08659 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# LoftQ Initialization\n\n| Paper | Code | PEFT Example |\n\nLoftQ (LoRA-fine-tuning-aware Quantization) provides a quantized backbone Q and LoRA adapters A and B, given a full-precision pre-trained weight W.\n\nThis model, 'Mistral-7B-v0.1-4bit-64rank', is obtained from Mistral-7B-v0.1. \nThe backbone is under 'LoftQ/Mistral-7B-v0.1-4bit-64rank' and LoRA adapters are under the 'subfolder='loftq_init''.## Model Info### Backbone\n- Stored format: 'torch.bfloat16'\n- Size: ~ 14 GiB\n- Loaded format: bitsandbytes nf4\n- Size loaded on GPU: ~3.5 GiB### LoRA adapters\n- rank: 64\n- lora_alpha: 16\n- target_modules: [\"down_proj\", \"up_proj\", \"q_proj\", \"k_proj\", \"v_proj\", \"o_proj\", \"gate_proj\"]## Usage\n\nTraining. Here's an example of loading this model and preparing for the LoRA fine-tuning.\n\n\n\nInference. Here is an example code for inference after the model has been fine-tuned on GSM8K.\n\n\n\nSee the full code at our Github Repo)" ]
[ -0.019296912476420403, 0.058481138199567795, -0.0029299550224095583, 0.08913679420948029, 0.10317570716142654, 0.027502182871103287, 0.08255332708358765, 0.14007462561130524, 0.015446123667061329, 0.11252390593290329, 0.05278438702225685, 0.09099206328392029, 0.11020182073116302, 0.16527986526489258, -0.011329758912324905, -0.20665962994098663, 0.0038374164141714573, -0.04317738115787506, 0.035542797297239304, 0.07518590986728668, 0.12384320795536041, -0.019196733832359314, 0.09108362346887589, -0.028890442103147507, -0.06895099580287933, -0.016004642471671104, -0.015839381143450737, -0.015046450309455395, 0.060150377452373505, 0.09890823811292648, 0.055793099105358124, -0.05026509240269661, 0.0677642971277237, -0.22091667354106903, 0.016042761504650116, 0.10858990252017975, 0.0033525677863508463, 0.0936117097735405, 0.049725133925676346, 0.007305649109184742, 0.05935653671622276, -0.15736791491508484, -0.012525057420134544, 0.05292237922549248, -0.08385401219129562, -0.17134889960289001, -0.11919286102056503, 0.04596269130706787, 0.06111884117126465, 0.053943078964948654, 0.01722855493426323, 0.11616621911525726, 0.04706288501620293, 0.06750205159187317, 0.17978516221046448, -0.22894421219825745, -0.08361022174358368, 0.15762461721897125, 0.02956942282617092, 0.10150735080242157, -0.07820647954940796, -0.029388543218374252, 0.02610655128955841, 0.024447321891784668, 0.00030033072107471526, -0.015484007075428963, -0.0428372323513031, 0.0038935947231948376, -0.16153134405612946, 0.0012804399011656642, 0.13804219663143158, 0.026249926537275314, -0.056977253407239914, -0.13983233273029327, -0.06004995480179787, -0.07859103381633759, -0.008373366668820381, 0.00464315852150321, -0.000516955740749836, 0.02713441662490368, 0.09010083228349686, -0.074836365878582, -0.052559323608875275, -0.0751456692814827, -0.0421864315867424, 0.08442651480436325, 0.008869672194123268, 0.061185289174318314, -0.016332989558577538, 0.09590797871351242, -0.1371728777885437, -0.11282037943601608, -0.05582108721137047, -0.08431726694107056, -0.0344923734664917, -0.03841738402843475, 0.007219625636935234, -0.02270800806581974, 0.07748356461524963, 0.2018701732158661, -0.07244207710027695, 0.07559402287006378, -0.0002406011917628348, 0.033672843128442764, -0.017090020701289177, 0.1668909341096878, -0.06535913795232773, -0.09302020072937012, 0.07544680684804916, 0.06169898062944412, 0.06811803579330444, -0.019849490374326706, -0.09224168956279755, -0.07732591032981873, -0.020453045144677162, 0.01792648434638977, 0.022136662155389786, 0.04189513623714447, -0.0856531411409378, -0.05220988392829895, 0.15939342975616455, -0.1214803010225296, 0.015393214300274849, -0.011415383778512478, -0.03562270104885101, 0.019934745505452156, 0.05312375724315643, -0.015021598897874355, -0.04507235437631607, -0.02759038284420967, -0.047473929822444916, -0.016633952036499977, -0.10266316682100296, -0.09154088795185089, 0.021169865503907204, -0.02877061627805233, -0.018292415887117386, -0.07443157583475113, -0.11029449850320816, -0.05225410684943199, 0.05889402702450752, -0.09377868473529816, 0.022562962025403976, -0.014485040679574013, -0.05779655650258064, 0.017092633992433548, 0.018578369170427322, 0.1031625047326088, -0.032446108758449554, 0.06293935328722, 0.00027329387376084924, 0.060273874551057816, -0.10171688348054886, -0.0005908756866119802, -0.054090797901153564, 0.07080543786287308, -0.16075269877910614, 0.09971338510513306, -0.07003392279148102, 0.030265063047409058, -0.1753711849451065, -0.019065264612436295, -0.04138175770640373, -0.003183390712365508, 0.07948162406682968, 0.059248339384794235, -0.1959095150232315, 0.016622424125671387, 0.10624697804450989, -0.08398709446191788, -0.019137196242809296, 0.09605872631072998, 0.0033665818627923727, 0.056788258254528046, 0.03194672241806984, 0.1385132223367691, 0.18731366097927094, -0.08302189409732819, -0.05032829940319061, 0.05405305698513985, -0.028492406010627747, -0.012502118945121765, 0.06332617998123169, 0.0006088006193749607, -0.08033517748117447, 0.0507170669734478, -0.0726652592420578, 0.04494756832718849, 0.02080335095524788, -0.028072092682123184, -0.05733685567975044, -0.02491358108818531, -0.0535128228366375, -0.012249556370079517, -0.056532520800828934, -0.008857417851686478, -0.1346624791622162, -0.07275370508432388, 0.11086337268352509, -0.022419165819883347, -0.003606359474360943, -0.05129142105579376, 0.07925873249769211, -0.02636217512190342, 0.01155233383178711, -0.11927834898233414, -0.03028031997382641, 0.030023958534002304, -0.07705394923686981, 0.005089165177196264, -0.0793604701757431, 0.07250235974788666, 0.04257526621222496, 0.0018747353460639715, -0.001736234757117927, 0.06731760501861572, -0.0442366786301136, -0.06911288946866989, -0.13238021731376648, -0.06077214330434799, -0.027857376262545586, 0.09280994534492493, -0.10463802516460419, 0.01902349293231964, 0.030433289706707, 0.13211004436016083, 0.012836530804634094, -0.044313620775938034, 0.021459728479385376, -0.019405610859394073, -0.006907190661877394, -0.10395936667919159, 0.029005544260144234, -0.010796772316098213, -0.030194895341992378, -0.017821239307522774, -0.11670736223459244, -0.07759914547204971, 0.10005632787942886, 0.10718835145235062, -0.04665277525782585, -0.10797259956598282, -0.008840175345540047, 0.004184985067695379, -0.051022015511989594, -0.0026591189671307802, 0.13932140171527863, 0.018937846645712852, 0.06650281697511673, -0.05010093003511429, -0.05585690587759018, 0.002416477305814624, 0.011044427752494812, -0.04552731290459633, 0.09402503818273544, 0.02035032957792282, -0.19724689424037933, 0.07476679980754852, 0.06816688925027847, -0.04766739159822464, 0.1748080551624298, -0.01926230825483799, -0.03288240730762482, -0.05495186522603035, 0.09635519236326218, 0.01312810555100441, 0.0784502848982811, -0.010373708792030811, 0.03817281872034073, 0.035225413739681244, -0.013699754141271114, 0.0020145440939813852, -0.10487794876098633, 0.022707071155309677, -0.02466522715985775, -0.05109405517578125, 0.026060068979859352, 0.01904131844639778, 0.004494132939726114, 0.06721419841051102, 0.047593265771865845, 0.05652261897921562, 0.02085549384355545, -0.04190785065293312, -0.05783962458372116, 0.157321035861969, -0.14332950115203857, -0.11931035667657852, -0.2215355783700943, -0.03765136003494263, -0.04118620976805687, -0.032964129000902176, -0.02170311100780964, -0.06563272327184677, -0.0740346759557724, -0.06479548662900925, 0.14151616394519806, 0.010301519185304642, 0.014394246973097324, -0.03180155158042908, 0.0410291962325573, 0.08804193139076233, -0.15623271465301514, 0.0020689107477664948, 0.02431967295706272, -0.09627489000558853, 0.02080920897424221, 0.0013717868132516742, 0.02343081496655941, 0.05084621533751488, 0.03215257450938225, 0.01663021184504032, -0.0007405790966004133, 0.30644652247428894, -0.06976491212844849, 0.11306743323802948, 0.15098921954631805, -0.0055379318073391914, 0.05947941169142723, 0.16578856110572815, 0.026714857667684555, -0.04740646854043007, 0.01245831698179245, 0.017931843176484108, -0.028227565810084343, -0.1897585541009903, -0.04096372798085213, -0.04666841775178909, -0.04064740240573883, 0.07687103003263474, 0.06967730820178986, -0.014152226969599724, 0.07628341019153595, -0.05658110976219177, 0.10611176490783691, 0.06475725769996643, 0.09069977700710297, 0.08739517629146576, 0.021094808354973793, 0.07277952879667282, -0.05953942984342575, 0.0072059230878949165, 0.09816774725914001, 0.08160161972045898, 0.15737475454807281, -0.07787938416004181, 0.1330656260251999, 0.006354366894811392, 0.15797369182109833, 0.05397786945104599, 0.050773102790117264, -0.054180681705474854, 0.019878191873431206, -0.01219892967492342, -0.07079722732305527, -0.06266782432794571, 0.057656820863485336, -0.05293133854866028, 0.07238443940877914, 0.04709719493985176, 0.0030347732827067375, 0.05752888694405556, 0.16980081796646118, 0.07037961483001709, -0.2500743269920349, -0.05515843257308006, 0.06357190012931824, 0.007130446843802929, -0.1175953671336174, 0.027794046327471733, 0.1326139122247696, -0.03463155776262283, 0.058873094618320465, -0.05897533521056175, 0.0707976445555687, -0.10740895569324493, -0.005476667545735836, 0.11937819421291351, 0.23506008088588715, -0.01427854597568512, 0.0762459859251976, -0.20246128737926483, 0.074419766664505, 0.05980854853987694, 0.04481790214776993, -0.07345998287200928, 0.07471302896738052, 0.01742871291935444, -0.03833245858550072, 0.12968860566616058, -0.006178874522447586, -0.04893582686781883, -0.19282743334770203, -0.1021692305803299, 0.009867113083600998, 0.10919835418462753, -0.03110424615442753, 0.09176261723041534, -0.08961726725101471, 0.0010441998019814491, -0.014443322084844112, -0.04567601904273033, -0.0644020140171051, -0.18692025542259216, 0.024344831705093384, 0.04719018191099167, -0.01741170324385166, -0.07145148515701294, 0.0038323896005749702, -0.016848504543304443, 0.07237634062767029, -0.1508965939283371, -0.08458101749420166, -0.12342497706413269, 0.057163361459970474, 0.1677132248878479, -0.0581873282790184, 0.017769426107406616, -0.04313813894987106, 0.09126250445842743, -0.016556140035390854, -0.15523689985275269, 0.06505409628152847, -0.04127022624015808, -0.10217583924531937, -0.02338668890297413, 0.11384221166372299, 0.018575746566057205, 0.008276877924799919, -0.010257061570882797, 0.04488816484808922, -0.03302378952503204, -0.12820877134799957, -0.0008246467332355678, 0.14175598323345184, -0.013357914984226227, 0.04930468276143074, -0.06454531848430634, -0.001993743935599923, -0.020648635923862457, 0.03568607568740845, 0.027975717559456825, 0.17719897627830505, -0.08591917157173157, 0.0712595209479332, -0.04620695486664772, -0.06846486032009125, -0.21114866435527802, -0.0072810351848602295, 0.09797327220439911, 0.03495869040489197, 0.006891459226608276, -0.10694867372512817, 0.11083143204450607, 0.08167748898267746, -0.020569773390889168, 0.13177205622196198, -0.1913977414369583, -0.0709935650229454, 0.04509172588586807, 0.07490210980176926, 0.022152191027998924, -0.13876336812973022, -0.057882994413375854, -0.017151663079857826, -0.09886402636766434, 0.08937085419893265, -0.11150568723678589, 0.06869971007108688, -0.03145598992705345, 0.03441941365599632, 0.03788478299975395, -0.03621220961213112, 0.14226754009723663, -0.08012336492538452, 0.03995422273874283, -0.06934060156345367, -0.008107510395348072, 0.02519574947655201, -0.07414164394140244, 0.05222573131322861, -0.06457439810037613, 0.05116325616836548, -0.11229857057332993, -0.022421851754188538, -0.021344805136322975, 0.031909093260765076, -0.029473084956407547, -0.040061503648757935, -0.062456998974084854, 0.0986044630408287, -0.010472091846168041, -0.020673632621765137, 0.11241039633750916, 0.002114952774718404, 0.07570691406726837, 0.17824439704418182, 0.05990083888173103, 0.044757358729839325, -0.007140427827835083, 0.06340134143829346, 0.00017944232968147844, 0.0679694265127182, -0.13678038120269775, 0.016986994072794914, 0.07979226857423782, -0.00101289723534137, 0.05367244407534599, 0.024347417056560516, -0.09036975353956223, -0.03752757981419563, 0.08596792817115784, -0.08524566143751144, -0.15019726753234863, -0.05452614277601242, 0.031498223543167114, -0.11514920741319656, 0.004188522230833769, 0.1550801694393158, 0.015224691480398178, -0.013329519890248775, 0.03389737755060196, 0.07253994047641754, -0.06606625765562057, 0.09627456218004227, 0.05613800510764122, 0.04593328386545181, -0.10478351265192032, 0.044505417346954346, 0.10696936398744583, 0.05982792004942894, 0.023768816143274307, 0.10627289116382599, -0.09049222618341446, -0.04680797830224037, 0.0223219133913517, 0.01981210708618164, 0.008853338658809662, -0.00990657601505518, 0.0192862655967474, -0.11143802106380463, -0.005262132268399, 0.014943561516702175, 0.0726570338010788, 0.005647343583405018, 0.005543677136301994, -0.026504501700401306, -0.08554726094007492, 0.09013400971889496, -0.0482943058013916, 0.05619487538933754, -0.1427392214536667, -0.054047707468271255, -0.03046399913728237, -0.021768804639577866, -0.017094209790229797, -0.0038721654564142227, -0.1875615119934082, -0.033043261617422104, -0.0925961285829544, 0.04583932086825371, -0.0471634604036808, 0.045949533581733704, 0.008793329820036888, -0.027401329949498177, -0.008418756537139416, 0.032224174588918686, -0.03074629232287407, -0.07425051182508469, -0.06144685298204422, 0.06155477836728096, -0.08804813772439957, -0.02889428846538067, 0.015933549031615257, -0.08200669288635254, 0.054722897708415985, 0.001060898881405592, 0.03363955765962601, 0.044970594346523285, -0.11837515234947205, 0.010812862776219845, 0.01838337816298008, 0.02306262031197548, 0.013343765400350094, -0.09950036555528641, 0.030485857278108597, -0.0275673009455204, -0.0325191505253315, -0.005838214419782162, 0.07815282046794891, -0.0913560539484024, 0.046588532626628876, -0.08588580787181854, 0.014589743688702583, -0.057008255273103714, 0.05785972252488136, 0.09895769506692886, 0.04865878447890282, 0.12162286043167114, -0.06915538758039474, 0.009790705516934395, -0.18374103307724, -0.012123563326895237, 0.005508379079401493, -0.05411108210682869, -0.07638496160507202, -0.024899868294596672, 0.06329132616519928, 0.0056784194894135, 0.11638598144054413, 0.02441500686109066, 0.015899179503321648, 0.011006824672222137, -0.007609611842781305, -0.02178758569061756, 0.0042653861455619335, 0.15498684346675873, 0.04364551231265068, 0.028847696259617805, 0.033006101846694946, 0.003430701792240143, 0.04714658483862877, 0.017994334921240807, 0.04444127902388573, 0.145304873585701, -0.032535672187805176, 0.03620104119181633, 0.05892639234662056, -0.09381823241710663, -0.016189340502023697, 0.2176443189382553, -0.03263526037335396, 0.014592276886105537, -0.08210182189941406, 0.06449505686759949, 0.08534451574087143, -0.1667778044939041, 0.026249080896377563, -0.0021858334075659513, -0.05068928003311157, -0.13817618787288666, -0.21511316299438477, -0.07401947677135468, -0.14113010466098785, -0.026606660336256027, -0.10430636256933212, 0.04242420569062233, 0.08664536476135254, 0.01470238994807005, 0.026200199499726295, 0.05039920657873154, 0.004556846339255571, -0.03743245080113411, 0.02118348516523838, 0.03762982413172722, 0.01066324207931757, 0.017302777618169785, -0.014165866188704967, 0.06940526515245438, -0.03034958988428116, 0.07017262279987335, 0.011446462012827396, 0.057680584490299225, 0.03496019169688225, -0.047586191445589066, -0.08357375115156174, -0.011888355016708374, 0.04252365604043007, -0.07703540474176407, 0.15260934829711914, 0.041440073400735855, -0.051323819905519485, -0.03093358315527439, 0.18933312594890594, -0.028334952890872955, -0.07937908172607422, -0.13920259475708008, 0.13487298786640167, -0.07700956612825394, 0.02019287645816803, -0.005150762852281332, -0.10168337821960449, 0.015328701585531235, 0.1804986298084259, 0.1570449322462082, -0.013913292437791824, -0.00019674129725899547, 0.054629847407341, -0.006556040607392788, -0.011545322835445404, 0.10278644412755966, 0.1175173968076706, 0.13269439339637756, -0.036106694489717484, 0.016446353867650032, 0.027233852073550224, -0.022003542631864548, -0.04658256843686104, 0.02821429818868637, -0.041576117277145386, -0.01403788197785616, -0.03547612950205803, 0.035967763513326645, -0.0492582842707634, -0.1298966407775879, 0.0628620833158493, -0.11071131378412247, -0.12216871231794357, -0.07817143201828003, 0.05194731429219246, -0.052651792764663696, 0.019724242389202118, -0.01364949345588684, -0.006444612517952919, 0.16949719190597534, -0.04095297306776047, -0.1289810836315155, -0.09991980344057083, 0.03306089714169502, 0.0002633423137012869, 0.1502375453710556, -0.003071726066991687, 0.03051474131643772, 0.1130928173661232, -0.012836307287216187, -0.16178607940673828, 0.0674542486667633, -0.0015721309464424849, -0.1267516165971756, 0.013933531939983368, 0.0955062285065651, -0.03792482241988182, 0.14537444710731506, 0.05307689309120178, -0.10317475348711014, 0.006404677871614695, -0.006921601481735706, -0.013994372449815273, -0.11393122375011444, 0.01378083135932684, -0.11635121703147888, 0.16946442425251007, 0.19206373393535614, -0.012716018594801426, 0.005601602140814066, -0.06779971718788147, 0.051258787512779236, -0.00928301177918911, 0.136163592338562, -0.009598806500434875, -0.08921802788972855, -0.024642406031489372, -0.07132554054260254, 0.01729740761220455, -0.16798442602157593, -0.05114324763417244, -0.05952920392155647, -0.02774052694439888, -0.062282245606184006, 0.11045130342245102, 0.1186642125248909, 0.03709263354539871, -0.04401509836316109, -0.07117228955030441, -0.019475260749459267, 0.07332568615674973, -0.08550611138343811, -0.04409599304199219 ]
null
null
transformers
# LoftQ Initialization | [Paper](https://arxiv.org/abs/2310.08659) | [Code](https://github.com/yxli2123/LoftQ) | [PEFT Example](https://github.com/huggingface/peft/tree/main/examples/loftq_finetuning) | LoftQ (LoRA-fine-tuning-aware Quantization) provides a quantized backbone Q and LoRA adapters A and B, given a full-precision pre-trained weight W. This model, `Llama-2-7b-hf-4bit-64rank`, is obtained from [LLAMA-2-7b](https://huggingface.co/meta-llama/Llama-2-7b-hf). The backbone is under `LoftQ/Llama-2-7b-hf-4bit-64rank` and LoRA adapters are under the `subfolder='loftq_init'`. ## Model Info ### Backbone - Stored format: `torch.bfloat16` - Size: ~ 14 GiB - Loaded format: bitsandbytes nf4 - Size loaded on GPU: ~3.5 GiB ### LoRA adapters - rank: 64 - lora_alpha: 16 - target_modules: ["down_proj", "up_proj", "q_proj", "k_proj", "v_proj", "o_proj", "gate_proj"] ## Usage **Training** Here's an example of loading this model and preparing for the LoRA fine-tuning. ```python import torch from transformers import AutoModelForCausalLM, BitsAndBytesConfig from peft import PeftModel MODEL_ID = "LoftQ/Llama-2-7b-hf-4bit-64rank" base_model = AutoModelForCausalLM.from_pretrained( MODEL_ID, torch_dtype=torch.bfloat16, # you may change it with different models quantization_config=BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_compute_dtype=torch.bfloat16, # bfloat16 is recommended bnb_4bit_use_double_quant=False, bnb_4bit_quant_type='nf4', ), ) peft_model = PeftModel.from_pretrained( base_model, MODEL_ID, subfolder="loftq_init", is_trainable=True, ) # Do training with peft_model ... ``` ## Experiment Results We have conducted experiments on supervised fine-tuning of [GSM8K](https://huggingface.co/datasets/gsm8k) and [WikiText-2](https://huggingface.co/datasets/wikitext/viewer/wikitext-2-raw-v1). | Model | Bits | Rank | LoRA Initial | GSM8K | WikiText-2 | | -------------- | ---- | ---- | -------------------- | ----- | ---------- | | LLAMA-2-7b | 16 | 64 | Gaussian + 0 | 36.9 | 5.08 | | LLAMA-2-7b | 4 | 64 | Gaussian + 0 (QLoRA) | 35.1 | 5.70 | | **LLAMA-2-7b** | 4 | 64 | LoftQ | 35.0 | 5.24 | **Inference** Here is an example code for inference after the model has been fine-tuned on [GSM8K](https://huggingface.co/datasets/gsm8k). ```python import torch from transformers import AutoModelForCausalLM, BitsAndBytesConfig from peft import PeftModel MODEL_ID = "LoftQ/Llama-2-7b-hf-4bit-64rank" base_model = AutoModelForCausalLM.from_pretrained( MODEL_ID, torch_dtype=torch.bfloat16, # you may change it with different models quantization_config=BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_compute_dtype=torch.bfloat16, # bfloat16 is recommended bnb_4bit_use_double_quant=False, bnb_4bit_quant_type='nf4', ), ) peft_model = PeftModel.from_pretrained( base_model, MODEL_ID, subfolder="gsm8k", is_trainable=True, ) # Do inference with peft_model ... ``` See the full code at our [Github Repo]((https://github.com/yxli2123/LoftQ)) ## Citation ```bibtex @article{li2023loftq, title={Loftq: Lora-fine-tuning-aware quantization for large language models}, author={Li, Yixiao and Yu, Yifan and Liang, Chen and He, Pengcheng and Karampatziakis, Nikos and Chen, Weizhu and Zhao, Tuo}, journal={arXiv preprint arXiv:2310.08659}, year={2023} } ```
{"language": ["en"], "license": "mit", "tags": ["quantization ", "lora"], "pipeline_tag": "text-generation"}
text-generation
ChuGyouk/Llama-2-7b-hf-4bit-64rank
[ "transformers", "safetensors", "llama", "text-generation", "quantization ", "lora", "en", "arxiv:2310.08659", "license:mit", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T15:35:34+00:00
[ "2310.08659" ]
[ "en" ]
TAGS #transformers #safetensors #llama #text-generation #quantization #lora #en #arxiv-2310.08659 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
LoftQ Initialization ==================== | Paper | Code | PEFT Example | LoftQ (LoRA-fine-tuning-aware Quantization) provides a quantized backbone Q and LoRA adapters A and B, given a full-precision pre-trained weight W. This model, 'Llama-2-7b-hf-4bit-64rank', is obtained from LLAMA-2-7b. The backbone is under 'LoftQ/Llama-2-7b-hf-4bit-64rank' and LoRA adapters are under the 'subfolder='loftq\_init''. Model Info ---------- ### Backbone * Stored format: 'torch.bfloat16' * Size: ~ 14 GiB * Loaded format: bitsandbytes nf4 * Size loaded on GPU: ~3.5 GiB ### LoRA adapters * rank: 64 * lora\_alpha: 16 * target\_modules: ["down\_proj", "up\_proj", "q\_proj", "k\_proj", "v\_proj", "o\_proj", "gate\_proj"] Usage ----- Training Here's an example of loading this model and preparing for the LoRA fine-tuning. Experiment Results ------------------ We have conducted experiments on supervised fine-tuning of GSM8K and WikiText-2. Inference Here is an example code for inference after the model has been fine-tuned on GSM8K. See the full code at our Github Repo)
[ "### Backbone\n\n\n* Stored format: 'torch.bfloat16'\n* Size: ~ 14 GiB\n* Loaded format: bitsandbytes nf4\n* Size loaded on GPU: ~3.5 GiB", "### LoRA adapters\n\n\n* rank: 64\n* lora\\_alpha: 16\n* target\\_modules: [\"down\\_proj\", \"up\\_proj\", \"q\\_proj\", \"k\\_proj\", \"v\\_proj\", \"o\\_proj\", \"gate\\_proj\"]\n\n\nUsage\n-----\n\n\nTraining Here's an example of loading this model and preparing for the LoRA fine-tuning.\n\n\nExperiment Results\n------------------\n\n\nWe have conducted experiments on supervised fine-tuning of GSM8K\nand WikiText-2.\n\n\n\nInference Here is an example code for inference after the model has been fine-tuned on GSM8K.\n\n\nSee the full code at our Github Repo)" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #quantization #lora #en #arxiv-2310.08659 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Backbone\n\n\n* Stored format: 'torch.bfloat16'\n* Size: ~ 14 GiB\n* Loaded format: bitsandbytes nf4\n* Size loaded on GPU: ~3.5 GiB", "### LoRA adapters\n\n\n* rank: 64\n* lora\\_alpha: 16\n* target\\_modules: [\"down\\_proj\", \"up\\_proj\", \"q\\_proj\", \"k\\_proj\", \"v\\_proj\", \"o\\_proj\", \"gate\\_proj\"]\n\n\nUsage\n-----\n\n\nTraining Here's an example of loading this model and preparing for the LoRA fine-tuning.\n\n\nExperiment Results\n------------------\n\n\nWe have conducted experiments on supervised fine-tuning of GSM8K\nand WikiText-2.\n\n\n\nInference Here is an example code for inference after the model has been fine-tuned on GSM8K.\n\n\nSee the full code at our Github Repo)" ]
[ 69, 49, 168 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #quantization #lora #en #arxiv-2310.08659 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Backbone\n\n\n* Stored format: 'torch.bfloat16'\n* Size: ~ 14 GiB\n* Loaded format: bitsandbytes nf4\n* Size loaded on GPU: ~3.5 GiB### LoRA adapters\n\n\n* rank: 64\n* lora\\_alpha: 16\n* target\\_modules: [\"down\\_proj\", \"up\\_proj\", \"q\\_proj\", \"k\\_proj\", \"v\\_proj\", \"o\\_proj\", \"gate\\_proj\"]\n\n\nUsage\n-----\n\n\nTraining Here's an example of loading this model and preparing for the LoRA fine-tuning.\n\n\nExperiment Results\n------------------\n\n\nWe have conducted experiments on supervised fine-tuning of GSM8K\nand WikiText-2.\n\n\n\nInference Here is an example code for inference after the model has been fine-tuned on GSM8K.\n\n\nSee the full code at our Github Repo)" ]
[ -0.018436850979924202, 0.04518907517194748, -0.0011448634322732687, 0.03661104664206505, 0.04303674399852753, -0.014646630734205246, 0.08046010136604309, 0.14851142466068268, 0.028619779273867607, 0.12422306835651398, 0.09787571430206299, 0.11499031633138657, 0.048931222409009933, 0.18761703372001648, -0.02168765850365162, -0.16294218599796295, 0.02168405055999756, -0.07957898825407028, -0.0024208719842135906, 0.1262126863002777, 0.11213807016611099, -0.04756546393036842, 0.10162270814180374, -0.0010730520589277148, -0.002196715911850333, -0.03160513564944267, -0.013311981223523617, -0.06792432069778442, 0.10993581265211105, 0.049326736479997635, 0.0077222902327775955, 0.03867240250110626, -0.022957831621170044, -0.22866779565811157, 0.01904814876616001, 0.05724700912833214, 0.046860065311193466, 0.06755688041448593, 0.06185312569141388, 0.02222716063261032, 0.17349901795387268, -0.08823796361684799, -0.033228322863578796, 0.06436628103256226, -0.11311101168394089, -0.1726938933134079, -0.14718054234981537, -0.024224326014518738, 0.08533191680908203, 0.05272791534662247, -0.05097098648548126, 0.09725142270326614, -0.04016854986548424, 0.06569328159093857, 0.2076064646244049, -0.3074920177459717, -0.028108613565564156, 0.1421196162700653, 0.07019373029470444, 0.04553401097655296, -0.05240628868341446, 0.03569198027253151, 0.07490447163581848, -0.034716419875621796, -0.046677712351083755, -0.031985923647880554, -0.03672753646969795, 0.027365613728761673, -0.16534502804279327, -0.05485163629055023, 0.1787158101797104, 0.0472479909658432, -0.09105291962623596, -0.1602611094713211, -0.0527496412396431, -0.15286628901958466, 0.012004071846604347, -0.02746582217514515, -0.004387151915580034, 0.01103116199374199, -0.01931476779282093, -0.06658472865819931, -0.0674232691526413, -0.11458994448184967, -0.019182030111551285, 0.19866682589054108, 0.05882440134882927, 0.01627926342189312, -0.03299756720662117, 0.16226361691951752, -0.15668724477291107, -0.10795551538467407, -0.12506306171417236, -0.05032939463853836, -0.021475164219737053, 0.0005958526162430644, -0.00409303791821003, 0.0024056497495621443, 0.1583089530467987, 0.29596349596977234, -0.04386170580983162, 0.031518787145614624, 0.02748066559433937, 0.020919986069202423, 0.0015052992384880781, 0.11995122581720352, -0.07954928278923035, -0.057412389665842056, 0.06878381222486496, 0.03721998259425163, 0.03393266722559929, -0.02126467600464821, -0.018387870863080025, -0.037824127823114395, 0.014677084051072598, 0.08851989358663559, 0.04720518738031387, 0.06270962208509445, -0.04537707567214966, -0.0006854418898001313, 0.0904853343963623, -0.10763628780841827, -0.05051137134432793, -0.024227682501077652, -0.037608321756124496, -0.09787996858358383, 0.01255773101001978, 0.03769364953041077, 0.019430993124842644, 0.020938970148563385, -0.05959518626332283, -0.06627111881971359, -0.06849376857280731, -0.054495830088853836, -0.004511675797402859, -0.00154182780534029, -0.001883092219941318, -0.11468232423067093, -0.1329348236322403, -0.06437575072050095, -0.013787666335701942, -0.06331482529640198, 0.00749572180211544, 0.008548700250685215, -0.06271518766880035, 0.001796411001123488, -0.03953230008482933, 0.07196389138698578, -0.07919144630432129, 0.10483883321285248, 0.08595550805330276, 0.07185157388448715, -0.011528116650879383, -0.02451127953827381, -0.1232588067650795, 0.09495929628610611, -0.1271258294582367, 0.08377519249916077, -0.13246050477027893, 0.007449367083609104, -0.17775185406208038, -0.04497666656970978, -0.047438111156225204, 0.009269796311855316, 0.08729717135429382, 0.11986428499221802, -0.12811985611915588, -0.022000104188919067, 0.17893429100513458, -0.08024179935455322, -0.10394704341888428, 0.13434258103370667, 0.016141364350914955, 0.05188120901584625, 0.04342103376984596, 0.1774958372116089, 0.043943390250205994, -0.0796540305018425, -0.09639598429203033, -0.02603963576257229, -0.009508403018116951, 0.06184285506606102, 0.12862178683280945, -0.06849750131368637, -0.04230324923992157, 0.039986368268728256, -0.10564211755990982, 0.030819332227110863, -0.05913802981376648, -0.04184717312455177, -0.03714900091290474, -0.0396452397108078, -0.006447374355047941, -0.04298010841012001, -0.019043007865548134, -0.09667612612247467, -0.13975182175636292, -0.2139025330543518, 0.1537763923406601, -0.01746116578578949, 0.00794981699436903, -0.11780022829771042, 0.1606721431016922, -0.0248472411185503, 0.015267468057572842, -0.08294916152954102, -0.02517358586192131, 0.03395048901438713, -0.03866470232605934, 0.04156230762600899, -0.03967341408133507, 0.049353085458278656, 0.04972316697239876, -0.01694301702082157, -0.011537279933691025, -0.014871915802359581, -0.02098453789949417, -0.07369208335876465, -0.10269396752119064, -0.039117682725191116, -0.009940208867192268, 0.11763783544301987, -0.11669369786977768, 0.005218085832893848, 0.13952355086803436, 0.08813708275556564, 0.03122381865978241, -0.068064846098423, 0.0445946604013443, -0.028139382600784302, -0.05533413216471672, -0.06272660195827484, 0.006458761636167765, -0.04322310537099838, -0.03215347230434418, 0.024660343304276466, -0.07221312820911407, -0.03554559499025345, 0.09217198193073273, 0.05851384624838829, -0.05330803990364075, -0.02834716998040676, -0.05719343200325966, 0.023328104987740517, -0.05536713823676109, -0.01242460124194622, 0.1168227568268776, 0.004856494255363941, 0.05671633780002594, -0.060251858085393906, -0.04835488274693489, -0.014714480377733707, -0.06798340380191803, 0.010481750592589378, 0.11206574738025665, 0.07519548386335373, -0.1129370853304863, 0.06141025573015213, 0.08980496972799301, -0.025420935824513435, 0.1695537567138672, -0.026973644271492958, -0.09767520427703857, -0.05050354078412056, 0.10555749386548996, 0.024225791916251183, 0.03462820127606392, -0.08573763072490692, 0.0239716824144125, 0.06835484504699707, 0.009014611132442951, 0.048680637031793594, -0.13760235905647278, -0.00951857678592205, 0.00840448122471571, -0.06881324201822281, 0.06724285334348679, 0.10017856955528259, -0.04453632980585098, 0.07993704080581665, 0.026076557114720345, 0.022436633706092834, 0.00276634912006557, -0.024972422048449516, -0.050758641213178635, 0.18460975587368011, -0.0661361888051033, -0.11010456830263138, -0.14963398873806, -0.04794515296816826, -0.06805245578289032, 0.00851100031286478, 0.06307564675807953, -0.1355533003807068, -0.08285606652498245, -0.05654026195406914, 0.12998613715171814, 0.07299663871526718, 0.02994951419532299, -0.024479249492287636, 0.027589039877057076, 0.0521574541926384, -0.15561315417289734, -0.016881003975868225, 0.014300277456641197, -0.14409169554710388, 0.06106162816286087, 0.04745280742645264, 0.09767807275056839, 0.1019563376903534, 0.03244701772928238, 0.02652951143682003, 0.007100573740899563, 0.29405614733695984, -0.09346236288547516, 0.006612223107367754, 0.24520565569400787, 0.03293633833527565, -0.007556585129350424, 0.09374409914016724, 0.00714071886613965, -0.13452275097370148, 0.07533026486635208, 0.07141762226819992, -0.05929065868258476, -0.2254754900932312, -0.043973784893751144, -0.033192843198776245, 0.05067010596394539, 0.009565887041389942, 0.08135136216878891, 0.05201499164104462, 0.09905564785003662, -0.05664477497339249, -0.005476413760334253, 0.02764006331562996, 0.0773504376411438, 0.05970044061541557, -0.019313044846057892, 0.08889757096767426, -0.041333843022584915, -0.038634128868579865, 0.10316430032253265, 0.05222397297620773, 0.09331053495407104, -0.03521302714943886, 0.15045183897018433, 0.11955208331346512, 0.12366307526826859, 0.05066416412591934, -0.0003545970539562404, -0.022253679111599922, -0.008772838860750198, -0.009295125491917133, -0.0871787890791893, -0.09455545246601105, 0.06633096933364868, -0.024470029398798943, -0.05657411366701126, 0.012010437436401844, 0.0398433580994606, 0.08103840798139572, 0.10209324210882187, 0.14844879508018494, -0.23407462239265442, -0.056324560195207596, 0.04130813106894493, -0.05423404276371002, -0.0673251673579216, 0.0801212266087532, 0.1190268024802208, -0.08579326421022415, 0.02159861847758293, -0.03843751549720764, 0.05887458473443985, -0.13214382529258728, 0.01597905531525612, 0.058941084891557693, 0.09682460129261017, -0.04186917096376419, 0.10625996440649033, -0.21715225279331207, 0.17420810461044312, 0.02331242710351944, 0.033449381589889526, -0.059928636997938156, 0.04907587543129921, 0.02716437168419361, -0.04570043459534645, 0.15684376657009125, -0.03470983728766441, -0.03761564940214157, -0.14472588896751404, -0.12801367044448853, 0.037977878004312515, 0.10143368691205978, -0.1294233798980713, 0.13710834085941315, -0.047466471791267395, 0.012762547470629215, -0.03920499607920647, 0.014321185648441315, -0.12911726534366608, -0.0923294797539711, 0.07412298768758774, 0.048655614256858826, 0.09167139232158661, -0.12627476453781128, -0.009259789250791073, -0.075055792927742, 0.17171710729599, -0.07202208787202835, -0.07100529968738556, -0.09721516072750092, 0.07026545703411102, 0.14175885915756226, -0.0933503806591034, 0.017051788046956062, -0.0037929804529994726, 0.03274243324995041, -0.002974448725581169, -0.05805227905511856, 0.09629500657320023, -0.05600416287779808, -0.2125484049320221, -0.003481407882645726, 0.15163500607013702, -0.015152920968830585, 0.022237639874219894, -0.008436791598796844, 0.05698211118578911, -0.013741987757384777, -0.13040784001350403, 0.060155611485242844, 0.04841771349310875, -0.028162533417344093, 0.09127277135848999, 0.04284682497382164, 0.049693237990140915, -0.06568536162376404, -0.07230986654758453, 0.12273282557725906, 0.35019221901893616, -0.028865111991763115, 0.03382168710231781, -0.0006954476120881736, -0.015586504712700844, -0.22311466932296753, -0.04204227775335312, 0.017928089946508408, 0.014193534851074219, -0.00011196227569598705, -0.12482354789972305, 0.06033075228333473, 0.11108506470918655, -0.06624209880828857, 0.12673960626125336, -0.27852919697761536, -0.16010555624961853, 0.09735028445720673, 0.061096515506505966, 0.10294216871261597, -0.1601017713546753, -0.09324824064970016, -0.0019099728669971228, -0.1792622059583664, 0.017318464815616608, -0.11981837451457977, 0.1275392770767212, -0.008411082439124584, 0.006516430992633104, 0.041357219219207764, -0.0651949942111969, 0.20733778178691864, -0.0452219694852829, 0.03288519009947777, -0.0732492133975029, -0.020268993452191353, -0.014079191721975803, -0.03449549525976181, 0.046557631343603134, -0.03330078348517418, 0.056608159095048904, -0.11127933859825134, -0.0539902001619339, -0.04043608903884888, 0.07808113843202591, -0.001477224170230329, -0.08741272240877151, -0.13754726946353912, 0.07554920017719269, -0.04065534099936485, 0.02074422501027584, 0.12371961772441864, -0.06689856201410294, 0.04317149519920349, 0.1376246213912964, 0.05901683121919632, 0.03146303817629814, -0.026055796071887016, 0.04904000461101532, 0.018019214272499084, 0.04980629310011864, -0.13347090780735016, 0.021088430657982826, 0.07421361654996872, 0.05483362823724747, 0.0676502138376236, 0.024614714086055756, -0.10527718812227249, 0.04726707562804222, 0.09850245714187622, -0.14910602569580078, -0.13683979213237762, -0.06930280476808548, -0.042255692183971405, -0.054687321186065674, 0.05181149020791054, 0.16355319321155548, 0.009846456348896027, -0.022694583982229233, -0.0002663700724951923, 0.09349525719881058, -0.07652778178453445, 0.18019933998584747, 0.046671561896800995, 0.06585774570703506, -0.1023247167468071, 0.024636918678879738, 0.07473769783973694, 0.029546599835157394, 0.03443581610918045, 0.03849092498421669, -0.07069504261016846, -0.038504112511873245, -0.0702677071094513, -0.007855750620365143, -0.02000785991549492, -0.03602147847414017, -0.05491361767053604, -0.06264762580394745, -0.013258174993097782, -0.025755595415830612, 0.08486989885568619, 0.06407522410154343, 0.030905021354556084, -0.005949666723608971, -0.08026981353759766, 0.1154065877199173, -0.001765272580087185, 0.08708956837654114, -0.1709252893924713, 0.03643076866865158, -0.05003826320171356, 0.007776573766022921, -0.02956925332546234, -0.0018273629248142242, -0.13310833275318146, -0.007368335034698248, -0.09512792527675629, -0.03493351489305496, -0.08516191691160202, -0.022093215957283974, 0.026360996067523956, -0.04834616929292679, -0.03885135054588318, 0.027098987251520157, -0.09238724410533905, -0.044129639863967896, -0.05440112203359604, 0.08943083137273788, -0.11159737408161163, 0.017203135415911674, 0.08105587214231491, -0.08112546801567078, 0.11519576609134674, -0.014733297750353813, 0.04098774865269661, 0.03789348155260086, -0.10764145106077194, 0.023705706000328064, 0.04371226951479912, -0.00045486257295124233, -0.020439567044377327, -0.1309680938720703, 0.022571153938770294, -0.04360451176762581, 0.0012295831693336368, 0.019440095871686935, 0.06470105051994324, -0.1381819099187851, -0.02219056338071823, -0.03610820323228836, 0.014940152876079082, -0.021511580795049667, 0.03169891610741615, 0.0553966760635376, 0.03378970921039581, 0.12307587265968323, -0.07274710386991501, -0.0053352368995547295, -0.1555250883102417, 0.013157479465007782, -0.005162699148058891, -0.07092586904764175, -0.04942447319626808, -0.010859671048820019, 0.07436957955360413, -0.049121350049972534, 0.16873253881931305, 0.014883507043123245, -0.07441538572311401, 0.006028319243341684, -0.037850793451070786, 0.1430360972881317, 0.012371435761451721, 0.23129905760288239, 0.05609096586704254, 0.032648131251335144, -0.03264223411679268, 0.014280545525252819, 0.0564071424305439, -0.09223996102809906, 0.17074821889400482, 0.1379542052745819, -0.043800488114356995, 0.037447333335876465, 0.021800948306918144, -0.13125549256801605, -0.045270659029483795, 0.07323835045099258, -0.05931459739804268, 0.06885112822055817, -0.04936002194881439, 0.03115122951567173, 0.1414833813905716, -0.14553089439868927, 0.01959207095205784, 0.001712186960503459, -0.029185988008975983, -0.149306982755661, -0.16248899698257446, -0.0840541422367096, -0.1790616512298584, 0.03818676248192787, -0.1122337132692337, 0.0511288158595562, 0.07378856092691422, 0.033474016934633255, -0.032332174479961395, 0.19413450360298157, -0.0965011864900589, -0.04588206112384796, 0.03016606904566288, -0.024128258228302002, -0.027871573343873024, -0.016164720058441162, -0.07126578688621521, 0.041887275874614716, 0.008479323238134384, 0.06840887665748596, -0.003540114499628544, -0.05560200288891792, 0.02381407842040062, -0.031030578538775444, -0.05220197141170502, -0.044932618737220764, 0.05308418720960617, 0.049260418862104416, 0.09466391056776047, 0.030749067664146423, -0.040012650191783905, 0.013872790150344372, 0.2434188723564148, -0.041969139128923416, -0.07354224473237991, -0.12136493623256683, 0.14544202387332916, -0.019463784992694855, 0.03885672986507416, 0.0019710268825292587, -0.12059711664915085, 0.01258345227688551, 0.20301419496536255, 0.14691753685474396, 0.011717062443494797, -0.01555461436510086, 0.02548839896917343, 0.00995919294655323, 0.007442532107234001, 0.06790874898433685, 0.07331140339374542, 0.17141962051391602, -0.0417654849588871, -0.042845338582992554, 0.04057189077138901, -0.02981993369758129, -0.025043638423085213, 0.08484607934951782, 0.010462160222232342, -0.05730975419282913, -0.039645999670028687, 0.06315144896507263, -0.07179277390241623, -0.12342855334281921, 0.0010488805128261447, -0.11187456548213959, -0.13885098695755005, -0.029802342876791954, 0.03342689573764801, 0.0022192765027284622, 0.047071099281311035, -0.014880266040563583, -0.03227695822715759, 0.039995886385440826, 0.005359731148928404, -0.10689081996679306, -0.04599523916840553, 0.06856318563222885, 0.014051620848476887, 0.22252371907234192, -0.004556841216981411, 0.06812959909439087, 0.13371224701404572, -0.07014521956443787, -0.14429356157779694, 0.1315343827009201, 0.038909465074539185, -0.13330820202827454, 0.04277490824460983, 0.11079870164394379, 0.02566547878086567, 0.07969924062490463, 0.07268397510051727, -0.15592510998249054, 0.014769644476473331, 0.08575733006000519, 0.003968218807131052, -0.1680489182472229, 0.023136869072914124, -0.15519820153713226, 0.15816648304462433, 0.149352565407753, -0.08151549845933914, 0.045067381113767624, -0.056464508175849915, 0.09934734553098679, 0.003034600056707859, 0.09693733602762222, -0.006970210932195187, -0.17106053233146667, -0.0038641414139419794, -0.03867500647902489, -0.016743360087275505, -0.2159852832555771, -0.021568812429904938, -0.017494944855570793, 0.023866068571805954, -0.05860951542854309, 0.11874305456876755, 0.18207386136054993, 0.013695682398974895, -0.02112528495490551, -0.11087316274642944, -0.02435768023133278, 0.11296574026346207, -0.15689943730831146, -0.06549379229545593 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text-generation
aidonuts/marigold-007-ep2
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T15:35:45+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 60, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.04654794931411743, 0.16618601977825165, -0.005445904564112425, 0.01853804849088192, 0.0981811136007309, 0.011998992413282394, 0.06433123350143433, 0.11398410052061081, -0.0230073444545269, 0.11406639218330383, 0.03047988750040531, 0.10172267258167267, 0.11317981779575348, 0.14841650426387787, -0.002152352826669812, -0.22403094172477722, 0.050844956189394, -0.12105348706245422, -0.033293843269348145, 0.11749980598688126, 0.1483822613954544, -0.09928343445062637, 0.07274559140205383, -0.029687678441405296, -0.012143402360379696, -0.030057786032557487, -0.05890674889087677, -0.046214159578084946, 0.04651786759495735, 0.06640566885471344, 0.06770290434360504, 0.0071083661168813705, 0.09012923389673233, -0.2696533799171448, 0.018959321081638336, 0.07145345956087112, -0.002759667346253991, 0.06957992166280746, 0.06404146552085876, -0.07107418030500412, 0.10337356477975845, -0.05106033384799957, 0.14650006592273712, 0.08365883678197861, -0.09081148356199265, -0.1895141303539276, -0.08866965025663376, 0.09882009029388428, 0.17572562396526337, 0.04925641790032387, -0.02320658043026924, 0.09761467576026917, -0.08769196271896362, 0.015438909642398357, 0.04981724172830582, -0.07620415836572647, -0.05378096550703049, 0.05986575037240982, 0.07907199114561081, 0.06627275794744492, -0.12434766441583633, -0.02885502204298973, 0.005009706597775221, 0.010980482213199139, 0.0769270583987236, 0.01728810742497444, 0.146672785282135, 0.0338633768260479, -0.12615777552127838, -0.04880760237574577, 0.09869225323200226, 0.03395522013306618, -0.04422314465045929, -0.24749068915843964, -0.03152675926685333, -0.030810698866844177, -0.029386121779680252, -0.03716538846492767, 0.04340358078479767, -0.007673026993870735, 0.08638741075992584, -0.0060646249912679195, -0.07403432577848434, -0.03937075287103653, 0.06169692054390907, 0.0672287791967392, 0.02999979443848133, -0.013745363801717758, 0.010938193649053574, 0.11620724946260452, 0.1095694974064827, -0.12054188549518585, -0.05555335059762001, -0.06393084675073624, -0.08656639605760574, -0.040790557861328125, 0.034162238240242004, 0.03456587344408035, 0.05349370837211609, 0.25305667519569397, 0.015654386952519417, 0.059652652591466904, 0.034477248787879944, 0.007892133668065071, 0.05848940089344978, 0.11044429242610931, -0.06018859148025513, -0.10444226115942001, -0.02648012898862362, 0.08843598514795303, 0.008199662901461124, -0.03287925571203232, -0.05088530853390694, 0.06019928678870201, 0.01946467161178589, 0.11926145106554031, 0.09061790257692337, 0.010536285117268562, -0.07121123373508453, -0.061038948595523834, 0.1891259253025055, -0.16544590890407562, 0.04322727024555206, 0.035097137093544006, -0.03903156518936157, 0.00019933005387429148, 0.013914269395172596, 0.016625655815005302, -0.025983380153775215, 0.09017423540353775, -0.054113563150167465, -0.04145489260554314, -0.11186197400093079, -0.03383193537592888, 0.033762916922569275, 0.008953776210546494, -0.035059962421655655, -0.033713940531015396, -0.08351044356822968, -0.07577689737081528, 0.09320491552352905, -0.07346344739198685, -0.04878907650709152, -0.01804324984550476, -0.07530532777309418, 0.022395428270101547, 0.019394835457205772, 0.07707412540912628, -0.02362251654267311, 0.04399976506829262, -0.05189276114106178, 0.05863580107688904, 0.11207318305969238, 0.03570080175995827, -0.05736649036407471, 0.06062258034944534, -0.23834340274333954, 0.09552820026874542, -0.07409077137708664, 0.05591456592082977, -0.153293639421463, -0.024439791217446327, 0.04788333550095558, 0.008784620091319084, -0.009650949388742447, 0.13416339457035065, -0.21702027320861816, -0.02536402828991413, 0.1717337965965271, -0.10057014971971512, -0.07069246470928192, 0.05619903281331062, -0.04835370555520058, 0.10988964140415192, 0.03825836628675461, -0.025690359994769096, 0.06171267107129097, -0.1267417073249817, 0.003717758459970355, -0.05005312338471413, -0.017048977315425873, 0.1548657864332199, 0.07182947546243668, -0.07217690348625183, 0.07399354875087738, 0.025708531960844994, -0.0246540866792202, -0.04625825211405754, -0.015164627693593502, -0.10536660254001617, 0.014689887873828411, -0.06369215250015259, 0.014470234513282776, -0.020807426422834396, -0.09071163833141327, -0.027962757274508476, -0.17504668235778809, -0.03014434315264225, 0.08651752024888992, -0.008693269453942776, -0.01803150773048401, -0.1178668737411499, 0.009341353550553322, 0.04177580401301384, 0.0061247628182172775, -0.13462838530540466, -0.04812471568584442, 0.02780051715672016, -0.1600649207830429, 0.034652888774871826, -0.05392369255423546, 0.04932025074958801, 0.025790516287088394, -0.028889117762446404, -0.026493212208151817, 0.021633783355355263, 0.005992184858769178, -0.011999987065792084, -0.24343903362751007, -0.028118690475821495, -0.024888472631573677, 0.1682123839855194, -0.20917098224163055, 0.03546025976538658, 0.07867541164159775, 0.15366052091121674, 0.011240328662097454, -0.04177491366863251, 0.005974748637527227, -0.06935794651508331, -0.02736494317650795, -0.05875484645366669, -0.0047869328409433365, -0.03310677409172058, -0.04545191675424576, 0.04568447172641754, -0.16510973870754242, -0.032636504620313644, 0.09776268899440765, 0.06289951503276825, -0.13922683894634247, -0.020621931180357933, -0.03630133345723152, -0.049253206700086594, -0.04911839962005615, -0.0605199858546257, 0.10893940925598145, 0.05891856551170349, 0.04574795812368393, -0.05928509309887886, -0.07568105310201645, -0.001827909960411489, -0.013898161239922047, -0.017864689230918884, 0.09759635478258133, 0.0751434788107872, -0.13251115381717682, 0.09224759042263031, 0.09603385627269745, 0.07919023185968399, 0.09113933145999908, -0.02355697751045227, -0.08261934667825699, -0.045987509191036224, 0.031442027539014816, 0.020124373957514763, 0.13039541244506836, -0.024294709786772728, 0.04352088272571564, 0.042134687304496765, -0.019369594752788544, 0.014752166345715523, -0.08687400817871094, 0.033972494304180145, 0.028472330421209335, -0.016721390187740326, 0.050190530717372894, -0.03876714035868645, 0.02440318465232849, 0.08830609917640686, 0.045322712510824203, 0.03507532551884651, 0.015493292361497879, -0.05206458270549774, -0.1083620935678482, 0.16405931115150452, -0.12714070081710815, -0.22483378648757935, -0.13936103880405426, 0.0037376401014626026, 0.035628627985715866, -0.015835661441087723, 0.002417160663753748, -0.059374887496232986, -0.12220635265111923, -0.08858037739992142, 0.015140829607844353, 0.04942670464515686, -0.09028962254524231, -0.06437795609235764, 0.058117836713790894, 0.03889724239706993, -0.14560972154140472, 0.017612040042877197, 0.04854894429445267, -0.09789852797985077, -0.006774199660867453, 0.08094939589500427, 0.0698540136218071, 0.1770169734954834, 0.017703235149383545, -0.021850809454917908, 0.032354529947042465, 0.20614571869373322, -0.13538233935832977, 0.11083246022462845, 0.13607586920261383, -0.09041404724121094, 0.08072979003190994, 0.19951270520687103, 0.03932560607790947, -0.10153959691524506, 0.031980328261852264, 0.02283124253153801, -0.0284719280898571, -0.24526868760585785, -0.07212468236684799, -0.004402178805321455, -0.058010730892419815, 0.07660572230815887, 0.09286724030971527, 0.08215958625078201, 0.012304253876209259, -0.09310996532440186, -0.08154371380805969, 0.05942574888467789, 0.10367169976234436, 0.024584239348769188, -0.010839897207915783, 0.08998730033636093, -0.034100502729415894, 0.019626356661319733, 0.0853661298751831, 0.005239574704319239, 0.17840281128883362, 0.05159219726920128, 0.18830420076847076, 0.07925192266702652, 0.07219027727842331, 0.009912233799695969, 0.013080619275569916, 0.018877580761909485, 0.03300119563937187, -0.002769160782918334, -0.08440786600112915, -0.02248465269804001, 0.11566436290740967, 0.06668911874294281, 0.010815348476171494, 0.015172341838479042, -0.04104290530085564, 0.07965951412916183, 0.1831512451171875, -0.007656289264559746, -0.1783534437417984, -0.057547420263290405, 0.07553383708000183, -0.09879875183105469, -0.09854305535554886, -0.013454320840537548, 0.03072015568614006, -0.17046253383159637, 0.023390959948301315, -0.02239842526614666, 0.1106182336807251, -0.14194999635219574, -0.020490378141403198, 0.07218493521213531, 0.07199500501155853, 0.004729843698441982, 0.05758659541606903, -0.16417601704597473, 0.10671813786029816, 0.008950476534664631, 0.06779605895280838, -0.09610627591609955, 0.1008887067437172, -0.004196076653897762, -0.02063460275530815, 0.1393408179283142, 0.002700034761801362, -0.06884108483791351, -0.0763031542301178, -0.08754398673772812, -0.009632662869989872, 0.12754282355308533, -0.1419651061296463, 0.08767123520374298, -0.037212442606687546, -0.0424150750041008, -0.0017086371080949903, -0.10206665843725204, -0.11638247221708298, -0.18888559937477112, 0.06001543253660202, -0.13492922484874725, 0.03152317553758621, -0.10799519717693329, -0.032371897250413895, -0.030304040759801865, 0.19337286055088043, -0.23447458446025848, -0.07199826091527939, -0.1475764364004135, -0.10233612358570099, 0.1443224400281906, -0.0501345656812191, 0.08485390990972519, -0.007241467013955116, 0.16846685111522675, 0.019060896709561348, -0.02531743235886097, 0.0971490666270256, -0.09173708409070969, -0.19302815198898315, -0.07869284600019455, 0.15662524104118347, 0.13260218501091003, 0.031680017709732056, -0.002461588243022561, 0.036563750356435776, -0.015421539545059204, -0.11935004591941833, 0.015969349071383476, 0.1787186712026596, 0.06237189099192619, 0.02331034652888775, -0.027346095070242882, -0.11273157596588135, -0.06900003552436829, -0.028530338779091835, 0.03054865077137947, 0.17762407660484314, -0.07057618349790573, 0.18207968771457672, 0.14163152873516083, -0.05922834202647209, -0.20400173962116241, 0.010538800619542599, 0.03055560030043125, 0.0009220078936778009, 0.02591954916715622, -0.20123432576656342, 0.08688826113939285, 0.004683020059019327, -0.05110127478837967, 0.13194532692432404, -0.17217805981636047, -0.14451217651367188, 0.0765485092997551, 0.038384392857551575, -0.19559739530086517, -0.12913893163204193, -0.09174312651157379, -0.045869920402765274, -0.18591414391994476, 0.09569250047206879, 0.0305706188082695, 0.010893458500504494, 0.03030681423842907, 0.029179483652114868, 0.019487828016281128, -0.0418255440890789, 0.18391458690166473, -0.024792250245809555, 0.026594700291752815, -0.08539514988660812, -0.06927408277988434, 0.03743394836783409, -0.052842434495687485, 0.07349982857704163, -0.023486759513616562, 0.007861839607357979, -0.10348054021596909, -0.042148489505052567, -0.03735732287168503, 0.015448716469109058, -0.09657872468233109, -0.08514349907636642, -0.045032672584056854, 0.09675803780555725, 0.09690850973129272, -0.033646680414676666, -0.028050623834133148, -0.07533035427331924, 0.04412057250738144, 0.19926515221595764, 0.1785389482975006, 0.042153384536504745, -0.08034496754407883, -0.004150947090238333, -0.010121207684278488, 0.04310847446322441, -0.20463712513446808, 0.06283636391162872, 0.05450061708688736, 0.01973269321024418, 0.11436162889003754, -0.019565396010875702, -0.15359151363372803, -0.07263088971376419, 0.06303015351295471, -0.060181066393852234, -0.19620554149150848, 0.00867035984992981, 0.060603946447372437, -0.16371412575244904, -0.04535605385899544, 0.04643881320953369, -0.005620351992547512, -0.038163937628269196, 0.021896906197071075, 0.09194854646921158, 0.0026654244866222143, 0.07427921891212463, 0.05387866869568825, 0.0827430784702301, -0.10537070035934448, 0.08090532571077347, 0.08839722722768784, -0.08452684432268143, 0.023530138656497, 0.10478579998016357, -0.059433579444885254, -0.03440561518073082, 0.020135708153247833, 0.08153781294822693, 0.01775863952934742, -0.040019966661930084, 0.013229827396571636, -0.10452935844659805, 0.05954122915863991, 0.08839859813451767, 0.032507482916116714, 0.016702456399798393, 0.03425082191824913, 0.04607953503727913, -0.07238735258579254, 0.12142276018857956, 0.031868141144514084, 0.017129309475421906, -0.036505792289972305, -0.040896978229284286, 0.019542274996638298, -0.03214648738503456, -0.005015232600271702, -0.03023446537554264, -0.07695909589529037, -0.014793801121413708, -0.1626158058643341, -0.011131818406283855, -0.05648450180888176, 0.010329355485737324, 0.03204665705561638, -0.032609567046165466, 0.008124498650431633, 0.009250079281628132, -0.07695289701223373, -0.0663459524512291, -0.020460480824112892, 0.09540658444166183, -0.16213038563728333, 0.022481130436062813, 0.08244425803422928, -0.12187694013118744, 0.09281346201896667, 0.016204802319407463, -0.006236857734620571, 0.025038830935955048, -0.1475188434123993, 0.034843120723962784, -0.03386561945080757, 0.010836300440132618, 0.04373383894562721, -0.21569781005382538, -0.00004886732858722098, -0.033673107624053955, -0.06639216095209122, -0.009451326914131641, -0.03672455996274948, -0.11508306115865707, 0.1058407872915268, 0.007236586883664131, -0.08753558248281479, -0.03186136856675148, 0.029325377196073532, 0.0838974118232727, -0.021959776058793068, 0.15145497024059296, -0.008370938710868359, 0.07429654151201248, -0.16209737956523895, -0.018623165786266327, -0.006028574425727129, 0.022658247500658035, -0.01664556935429573, -0.01111356820911169, 0.044031109660863876, -0.022746501490473747, 0.17925859987735748, -0.030318550765514374, 0.02272745408117771, 0.06815794110298157, 0.019072026014328003, -0.030184008181095123, 0.10406795144081116, 0.04094860330224037, 0.02014910988509655, 0.018591465428471565, 0.003289656015112996, -0.04647882282733917, -0.03173251822590828, -0.19407226145267487, 0.07288651913404465, 0.15608493983745575, 0.09729263186454773, -0.016707008704543114, 0.07954329252243042, -0.10199416428804398, -0.1109243705868721, 0.12477338314056396, -0.04797708988189697, -0.002418199321255088, -0.07150927931070328, 0.13247236609458923, 0.1437523066997528, -0.1859612911939621, 0.07269313186407089, -0.0699717253446579, -0.04708027467131615, -0.10980689525604248, -0.19441905617713928, -0.05561789125204086, -0.049456022679805756, -0.016053348779678345, -0.04698808491230011, 0.07504211366176605, 0.054538097232580185, 0.006766852922737598, -0.0023397188633680344, 0.06506035476922989, -0.031050674617290497, -0.0037882844917476177, 0.032597362995147705, 0.06591679900884628, 0.012734474614262581, -0.030802709981799126, 0.016619903966784477, -0.013545602560043335, 0.045626189559698105, 0.06578011065721512, 0.04976864159107208, -0.02938537672162056, 0.014603170566260815, -0.038539156317710876, -0.10249634087085724, 0.043612558394670486, -0.024421939626336098, -0.0789753645658493, 0.15477414429187775, 0.023680059239268303, 0.007779473438858986, -0.020137663930654526, 0.23901568353176117, -0.0738423764705658, -0.0964353010058403, -0.14737580716609955, 0.10557299107313156, -0.038081806153059006, 0.05800395458936691, 0.04625935107469559, -0.10226529091596603, 0.018044332042336464, 0.1338089406490326, 0.16182038187980652, -0.039008259773254395, 0.020095856860280037, 0.031135575845837593, 0.00566398398950696, -0.03622615709900856, 0.04847532883286476, 0.06906453520059586, 0.16569648683071136, -0.04632584750652313, 0.09100406616926193, 0.0019041687482967973, -0.09579581767320633, -0.038361791521310806, 0.11069868505001068, -0.016052277758717537, 0.019335128366947174, -0.05818064883351326, 0.11742528527975082, -0.06386786699295044, -0.23783175647258759, 0.06453443318605423, -0.0684293657541275, -0.13765870034694672, -0.02378307841718197, 0.08207765966653824, -0.012955902144312859, 0.027587108314037323, 0.0730307325720787, -0.07240920513868332, 0.201939657330513, 0.03798431158065796, -0.05499868467450142, -0.055047210305929184, 0.0805421993136406, -0.10008571296930313, 0.2739645540714264, 0.01557221356779337, 0.04601577669382095, 0.10384146869182587, -0.009341772645711899, -0.13838784396648407, 0.019836371764540672, 0.09581108391284943, -0.10502193123102188, 0.04196618124842644, 0.19815568625926971, -0.0014755994779989123, 0.12389086186885834, 0.07657600939273834, -0.07551808655261993, 0.0478031262755394, -0.08054235577583313, -0.06760486960411072, -0.09260394424200058, 0.09703279286623001, -0.07772123068571091, 0.14251399040222168, 0.13876807689666748, -0.05074559152126312, 0.012724342755973339, -0.031311117112636566, 0.044293127954006195, -0.00010600237874314189, 0.10321761667728424, 0.004272161517292261, -0.1832672357559204, 0.024692710489034653, 0.005650998093187809, 0.10749758034944534, -0.16033467650413513, -0.09566054493188858, 0.042343202978372574, 0.003505636239424348, -0.0672195628285408, 0.1290110945701599, 0.05665452033281326, 0.04342988133430481, -0.03997718170285225, -0.03521440550684929, -0.0060732318088412285, 0.13561366498470306, -0.10713256150484085, 0.0009933578548952937 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text-generation
Novin-AI/MeduWen-S
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-10T15:35:51+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #qwen2 #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #qwen2 #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 52, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #qwen2 #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.06744060665369034, 0.1237388551235199, -0.004114609677344561, 0.02991606667637825, 0.11460870504379272, 0.005568372085690498, 0.06294357031583786, 0.10971193760633469, -0.026014693081378937, 0.11581014841794968, 0.018924949690699577, 0.10499268025159836, 0.10659246146678925, 0.1691424399614334, -0.006015846040099859, -0.21231532096862793, 0.044865865260362625, -0.13380737602710724, -0.025073938071727753, 0.11961860954761505, 0.13043774664402008, -0.12202122807502747, 0.06986955553293228, -0.03994565084576607, -0.009295043535530567, -0.0361013263463974, -0.05820033326745033, -0.04808541759848595, 0.06927672773599625, 0.0690578892827034, 0.06336662918329239, 0.01922842301428318, 0.10299910604953766, -0.2810887396335602, 0.0236574187874794, 0.08111110329627991, 0.002226806478574872, 0.07000467926263809, 0.06337219476699829, -0.07296913117170334, 0.06984713673591614, -0.06522127240896225, 0.14495620131492615, 0.08224987238645554, -0.0922221839427948, -0.19323916733264923, -0.08794740587472916, 0.09357348084449768, 0.19385994970798492, 0.05913294479250908, -0.03049401193857193, 0.12686537206172943, -0.07434657961130142, 0.01852177456021309, 0.06567037850618362, -0.08194528520107269, -0.053086262196302414, 0.06812959164381027, 0.07113085687160492, 0.10160701721906662, -0.13397133350372314, -0.0072817932814359665, 0.03036416508257389, 0.013016993179917336, 0.10258961468935013, 0.017448842525482178, 0.11838137358427048, 0.04335033521056175, -0.14493173360824585, -0.038016412407159805, 0.0884561613202095, 0.04341543838381767, -0.05371417850255966, -0.24333322048187256, -0.021258622407913208, -0.033045537769794464, -0.03133222460746765, -0.048901937901973724, 0.046065423637628555, -0.018345197662711143, 0.0746571272611618, -0.00905180536210537, -0.077952079474926, -0.047369781881570816, 0.07820919156074524, 0.06576532125473022, 0.026357414200901985, -0.0243342574685812, 0.00772935152053833, 0.11627262830734253, 0.09934048354625702, -0.11843404918909073, -0.049750957638025284, -0.06367483735084534, -0.08425901085138321, -0.04867105185985565, 0.029223250225186348, 0.03197961300611496, 0.05072800815105438, 0.2138856053352356, -0.0016585314879193902, 0.04777570813894272, 0.0300018060952425, 0.01629858836531639, 0.0634123831987381, 0.09685925394296646, -0.058943528681993484, -0.12131623923778534, -0.022760409861803055, 0.10975006967782974, 0.002361652674153447, -0.03354809433221817, -0.04929806664586067, 0.0689367800951004, 0.017635801807045937, 0.12228328734636307, 0.07093650102615356, 0.01461301650851965, -0.07341200113296509, -0.0643758624792099, 0.17208924889564514, -0.1599913388490677, 0.033031485974788666, 0.027699848636984825, -0.049781136214733124, -0.016962584108114243, 0.0206128042191267, 0.030544809997081757, -0.009477566927671432, 0.08983151614665985, -0.051631052047014236, -0.03264494985342026, -0.11271350830793381, -0.05229318514466286, 0.022805018350481987, 0.02329850196838379, -0.029599839821457863, -0.04297630116343498, -0.10461901128292084, -0.0702618658542633, 0.08274642378091812, -0.06679617613554001, -0.04588131234049797, -0.034392643719911575, -0.08036767691373825, 0.012772615067660809, 0.006944936700165272, 0.11524419486522675, -0.024861354380846024, 0.04965236783027649, -0.05080482363700867, 0.07076980918645859, 0.12968726456165314, 0.0256124809384346, -0.052786268293857574, 0.05227842554450035, -0.23543758690357208, 0.10626004636287689, -0.07104437053203583, 0.04600486531853676, -0.16222067177295685, -0.019692296162247658, 0.04013443738222122, 0.022423196583986282, -0.0052419379353523254, 0.13304713368415833, -0.20579689741134644, -0.03484721481800079, 0.1778334081172943, -0.10716996341943741, -0.08844240009784698, 0.05829978361725807, -0.05727203190326691, 0.12106184661388397, 0.046658918261528015, -0.015959804877638817, 0.030861597508192062, -0.14105893671512604, -0.012573265470564365, -0.05725134164094925, -0.027953004464507103, 0.1594742387533188, 0.06174226105213165, -0.04975385218858719, 0.06329082697629929, 0.017857130616903305, -0.014720242470502853, -0.047373462468385696, -0.03508519008755684, -0.10099945217370987, 0.009225212968885899, -0.0735674798488617, 0.025139320641756058, -0.03237168863415718, -0.09091918170452118, -0.030487151816487312, -0.15721407532691956, 0.006027343682944775, 0.09086263924837112, -0.0028123122174292803, -0.02166888304054737, -0.10495693236589432, -0.015849687159061432, 0.023717699572443962, 0.0010735627729445696, -0.14732947945594788, -0.052729055285453796, 0.01963592879474163, -0.16102278232574463, 0.03527507185935974, -0.032337408512830734, 0.046559423208236694, 0.04404491186141968, -0.044810350984334946, -0.03644292429089546, 0.01527401339262724, 0.01702694222331047, -0.01812152937054634, -0.2757890224456787, -0.016599029302597046, -0.037502363324165344, 0.16484688222408295, -0.2536672055721283, 0.044451385736465454, 0.052858345210552216, 0.12650004029273987, 0.011718528345227242, -0.026840604841709137, 0.02031077817082405, -0.06778053194284439, -0.03378141298890114, -0.060537584125995636, -0.0102090397849679, -0.036261335015296936, -0.05234677344560623, 0.03442572429776192, -0.16672758758068085, -0.04233158379793167, 0.11038065701723099, 0.03841483220458031, -0.1514066904783249, -0.046796903014183044, -0.04655757546424866, -0.05544671788811684, -0.06981822848320007, -0.05111313611268997, 0.10990618914365768, 0.0552663654088974, 0.054820816963911057, -0.06279280036687851, -0.06714518368244171, 0.008098754100501537, -0.023038236424326897, -0.01628015749156475, 0.08303935825824738, 0.07147926092147827, -0.12255207449197769, 0.09013188630342484, 0.0958702489733696, 0.08535332977771759, 0.10111390799283981, 0.0031223141122609377, -0.08790350705385208, -0.02990630455315113, 0.029989181086421013, 0.01356097124516964, 0.150030717253685, -0.026905570179224014, 0.049839962273836136, 0.03979787230491638, -0.007262712344527245, 0.005843297578394413, -0.0978906974196434, 0.029100263491272926, 0.024840185418725014, -0.011728756129741669, 0.036994971334934235, -0.05755846947431564, 0.016809193417429924, 0.10532841086387634, 0.040135741233825684, 0.051635969430208206, 0.008006487041711807, -0.05116545408964157, -0.11712050437927246, 0.1763288974761963, -0.11831972748041153, -0.23028700053691864, -0.12128487974405289, -0.012982514686882496, 0.03150848671793938, -0.012953351251780987, 0.025938911363482475, -0.07433073222637177, -0.11664986610412598, -0.0922725722193718, 0.04694730415940285, 0.059740062803030014, -0.08346977084875107, -0.062362488359212875, 0.06679393351078033, 0.0457296296954155, -0.1380528211593628, 0.026153815910220146, 0.035679563879966736, -0.09117627143859863, 0.005887721199542284, 0.08140957355499268, 0.06103856489062309, 0.1818755865097046, 0.012728521600365639, -0.023938871920108795, 0.019584620371460915, 0.20903365314006805, -0.136505126953125, 0.10589402914047241, 0.13493265211582184, -0.0703483521938324, 0.08147261291742325, 0.2107224464416504, 0.0418342649936676, -0.10617547482252121, 0.04455582797527313, 0.034235551953315735, -0.0238803718239069, -0.25054290890693665, -0.07808786630630493, 0.007576430216431618, -0.06175751984119415, 0.06809944659471512, 0.08130444586277008, 0.09570267051458359, 0.01984638161957264, -0.10488120466470718, -0.06586658954620361, 0.05113326013088226, 0.11108365654945374, -0.007418854162096977, -0.012006757780909538, 0.0969165563583374, -0.020286425948143005, 0.028002621605992317, 0.09235991537570953, 0.0084880031645298, 0.18746548891067505, 0.05100390687584877, 0.14692288637161255, 0.09142749756574631, 0.06584213674068451, 0.015684716403484344, 0.006666323635727167, 0.015644695609807968, 0.02073444239795208, -0.014378254301846027, -0.0880797803401947, -0.0017288135131821036, 0.12815876305103302, 0.020411469042301178, 0.050393857061862946, 0.005088018253445625, -0.032580070197582245, 0.08683152496814728, 0.17358696460723877, 0.010363306850194931, -0.1908130794763565, -0.07101033627986908, 0.06939493864774704, -0.08181700855493546, -0.10146915167570114, -0.02635601907968521, 0.04305123910307884, -0.17831183969974518, 0.014033086597919464, -0.022382382303476334, 0.10410568863153458, -0.11462701857089996, -0.012489398010075092, 0.04906824603676796, 0.07298072427511215, -0.016658522188663483, 0.06773389875888824, -0.18002092838287354, 0.1395270675420761, 0.01758507452905178, 0.07150158286094666, -0.08825206011533737, 0.08410486578941345, 0.003178939688950777, 0.0013509939890354872, 0.14415407180786133, 0.0013785995543003082, -0.0523817352950573, -0.10979107022285461, -0.08634650707244873, -0.009079654701054096, 0.13044366240501404, -0.12778301537036896, 0.10016698390245438, -0.01834736578166485, -0.045373477041721344, 0.005183245521038771, -0.11240560561418533, -0.14056962728500366, -0.1725207269191742, 0.04330243170261383, -0.13124029338359833, 0.04465160518884659, -0.10545487701892853, -0.048093315213918686, -0.05306214094161987, 0.19742146134376526, -0.22286871075630188, -0.07013117522001266, -0.1519971340894699, -0.05761480703949928, 0.119932159781456, -0.04775578901171684, 0.08312731981277466, 0.012994625605642796, 0.18674440681934357, 0.014313536696135998, -0.013770169578492641, 0.11090241372585297, -0.10466983169317245, -0.21406547725200653, -0.10291838645935059, 0.14246919751167297, 0.13924811780452728, 0.041273895651102066, 0.0022257522214204073, 0.02827414683997631, -0.014804026111960411, -0.11688549816608429, 0.020713498815894127, 0.1711113303899765, 0.11356078088283539, 0.031762681901454926, -0.045852549374103546, -0.12838490307331085, -0.08528922498226166, -0.04527286812663078, 0.01937401480972767, 0.1929924041032791, -0.07334718853235245, 0.17354312539100647, 0.15734395384788513, -0.05666225776076317, -0.1967383325099945, 0.02808118239045143, 0.04254651814699173, 0.0018926940392702818, 0.058352239429950714, -0.19716250896453857, 0.0960150957107544, 0.0021078127902001143, -0.054582200944423676, 0.11626559495925903, -0.18086016178131104, -0.1472223997116089, 0.055250246077775955, 0.06544214487075806, -0.1867036670446396, -0.12468403577804565, -0.09152166545391083, -0.040479280054569244, -0.12750375270843506, 0.08364081382751465, -0.015219016931951046, 0.011511581018567085, 0.03329310938715935, 0.02034589648246765, 0.010542148724198341, -0.043612707406282425, 0.18297483026981354, -0.0074994368478655815, 0.04291056841611862, -0.07745802402496338, -0.06123793497681618, 0.04548247158527374, -0.06682101637125015, 0.0688505694270134, -0.012457388453185558, 0.01576600968837738, -0.10679414868354797, -0.05470338836312294, -0.03223368898034096, 0.019370099529623985, -0.08504306524991989, -0.10194364190101624, -0.036353081464767456, 0.09871356934309006, 0.09517461061477661, -0.037792425602674484, -0.056679584085941315, -0.08485732227563858, 0.04062115028500557, 0.20317383110523224, 0.18020522594451904, 0.053560756146907806, -0.06437430530786514, -0.006059312727302313, -0.013237647712230682, 0.049002740532159805, -0.22129850089550018, 0.05923459306359291, 0.041168149560689926, 0.03180031478404999, 0.11860810965299606, -0.023935925215482712, -0.1587793081998825, -0.0502057746052742, 0.05410148575901985, -0.07425004243850708, -0.1685684472322464, 0.010434879921376705, 0.08286356180906296, -0.1552492380142212, -0.022906674072146416, 0.04575012996792793, -0.020043641328811646, -0.03438226878643036, 0.00707294000312686, 0.07919111847877502, 0.009836919605731964, 0.08478374034166336, 0.057017721235752106, 0.0959276556968689, -0.10216023027896881, 0.06617968529462814, 0.08096546679735184, -0.09338610619306564, 0.03410530090332031, 0.07545924931764603, -0.07126593589782715, -0.037233464419841766, 0.04482624679803848, 0.0918767899274826, 0.031775590032339096, -0.050642579793930054, 0.012327476404607296, -0.10012588649988174, 0.05418751388788223, 0.11697539687156677, 0.03980601206421852, 0.0020653458777815104, 0.0349934883415699, 0.04598642885684967, -0.09361135214567184, 0.12619003653526306, 0.03253564611077309, 0.024358928203582764, -0.044029660522937775, -0.027948984876275063, 0.033686719834804535, -0.020634718239307404, -0.014900618232786655, -0.04131974279880524, -0.06906769424676895, -0.011919837445020676, -0.17663416266441345, -0.0006877299747429788, -0.03835081309080124, 0.008035878650844097, 0.01438689511269331, -0.03798643499612808, 0.008271864615380764, 0.015990857034921646, -0.07275852560997009, -0.05440134555101395, -0.01070401445031166, 0.10120883584022522, -0.16839949786663055, 0.013798215426504612, 0.0738481730222702, -0.11845122277736664, 0.08829576522111893, 0.01660950295627117, 0.004566526506096125, 0.03947852551937103, -0.12990154325962067, 0.0469437912106514, -0.015183643437922001, 0.017251212149858475, 0.051821283996105194, -0.20713716745376587, -0.005219681188464165, -0.053738780319690704, -0.054747533053159714, -0.008454185910522938, -0.028378764167428017, -0.11614704132080078, 0.10657370090484619, 0.006339828949421644, -0.07519937306642532, -0.027563083916902542, 0.034499529749155045, 0.07487460225820541, -0.031029552221298218, 0.1542745679616928, -0.014918236993253231, 0.06987065821886063, -0.1874280571937561, -0.023337583988904953, -0.014252493157982826, 0.024976249784231186, -0.03739270567893982, -0.01777520589530468, 0.05066380277276039, -0.025644395500421524, 0.1947220265865326, -0.02277233451604843, 0.05517526715993881, 0.06517178565263748, -0.015353423543274403, -0.025753356516361237, 0.10341554135084152, 0.055761225521564484, 0.015996338799595833, 0.03251899033784866, 0.007716674357652664, -0.03165765851736069, -0.005552713759243488, -0.167100191116333, 0.07967466861009598, 0.16496649384498596, 0.08635497838258743, -0.014588052406907082, 0.06132662668824196, -0.11290588229894638, -0.11605644226074219, 0.09777160733938217, -0.056159622967243195, -0.01740921474993229, -0.062441661953926086, 0.13894620537757874, 0.1522199958562851, -0.19082458317279816, 0.06211152300238609, -0.06795507669448853, -0.0487544871866703, -0.10746019333600998, -0.16687791049480438, -0.05764069780707359, -0.05954143404960632, -0.020104030147194862, -0.05745544657111168, 0.06959457695484161, 0.07283110171556473, 0.017621422186493874, 0.012575851753354073, 0.07775423675775528, -0.017673097550868988, 0.00843984168022871, 0.026977673172950745, 0.06567810475826263, 0.013495570048689842, -0.04381807893514633, 0.016235843300819397, -0.00015613723371643573, 0.034048307687044144, 0.047009509056806564, 0.039173372089862823, -0.03012777306139469, 0.005396591499447823, -0.03004968911409378, -0.1132737398147583, 0.04056783393025398, -0.0245139729231596, -0.06442589312791824, 0.13803128898143768, 0.026449358090758324, -0.006702050566673279, -0.025474393740296364, 0.2641041576862335, -0.07600386440753937, -0.09474562108516693, -0.13578693568706512, 0.13365262746810913, -0.0308542363345623, 0.06413768976926804, 0.033664409071207047, -0.11381697654724121, 0.027896301820874214, 0.145524263381958, 0.14766931533813477, -0.059594202786684036, 0.018058648332953453, 0.023248950019478798, 0.0036677704192698, -0.038663145154714584, 0.05093686655163765, 0.07642526924610138, 0.13084270060062408, -0.057510439306497574, 0.07993458956480026, -0.00528855761513114, -0.09648048877716064, -0.03070426546037197, 0.12046385556459427, -0.005974611733108759, 0.018961863592267036, -0.06711561232805252, 0.12644343078136444, -0.043718259781599045, -0.261628121137619, 0.05282887443900108, -0.06905496120452881, -0.14716462790966034, -0.02855629473924637, 0.05909299477934837, -0.00726199010387063, 0.02540661208331585, 0.06713409721851349, -0.06904488801956177, 0.19428247213363647, 0.03470597416162491, -0.044902503490448, -0.06258992105722427, 0.07463990896940231, -0.10928831994533539, 0.28889188170433044, 0.010627356357872486, 0.05702703818678856, 0.1010323017835617, -0.02710605598986149, -0.13230937719345093, 0.030603965744376183, 0.08569987118244171, -0.08157077431678772, 0.049359869211912155, 0.2173999398946762, -0.00799210648983717, 0.11221332848072052, 0.0741662085056305, -0.09916665405035019, 0.052276816219091415, -0.10220054537057877, -0.09391136467456818, -0.08265925943851471, 0.09803684055805206, -0.05557653307914734, 0.14824360609054565, 0.12248145043849945, -0.04785078391432762, 0.022196060046553612, -0.022353654727339745, 0.04894673451781273, 0.006722010672092438, 0.12958186864852905, 0.013888917863368988, -0.19708466529846191, 0.027539461851119995, -0.004416270647197962, 0.09896787256002426, -0.2124645709991455, -0.10066045075654984, 0.05214649438858032, 0.00458158552646637, -0.06152847036719322, 0.12505200505256653, 0.06458623707294464, 0.040626320987939835, -0.045448239892721176, -0.0330616720020771, -0.008380461484193802, 0.1610291600227356, -0.10901795327663422, -0.004472559317946434 ]
null
null
transformers
# LoftQ Initialization | [Paper](https://arxiv.org/abs/2310.08659) | [Code](https://github.com/yxli2123/LoftQ) | [PEFT Example](https://github.com/huggingface/peft/tree/main/examples/loftq_finetuning) | LoftQ (LoRA-fine-tuning-aware Quantization) provides a quantized backbone Q and LoRA adapters A and B, given a full-precision pre-trained weight W. This model, `Llama-2-13b-hf-4bit-64rank`, is obtained from [LLAMA-2-13b](https://huggingface.co/meta-llama/Llama-2-13b-hf). The backbone is under `LoftQ/Llama-2-13b-hf-4bit-64rank` and LoRA adapters are under the `subfolder='loftq_init'`. ## Model Info ### Backbone - Stored format: `torch.bfloat16` - Size: ~ 26 GiB - Loaded format: bitsandbytes nf4 - Size loaded on GPU: ~6.5 GiB ### LoRA adapters - rank: 64 - lora_alpha: 64 - target_modules: ["down_proj", "up_proj", "q_proj", "k_proj", "v_proj", "o_proj", "gate_proj"] ## Usage **Training** Here's an example of loading this model and preparing for the LoRA fine-tuning. ```python import torch from transformers import AutoModelForCausalLM, BitsAndBytesConfig from peft import PeftModel MODEL_ID = "LoftQ/Llama-2-13b-hf-4bit-64rank" base_model = AutoModelForCausalLM.from_pretrained( MODEL_ID, torch_dtype=torch.bfloat16, # you may change it with different models quantization_config=BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_compute_dtype=torch.bfloat16, # bfloat16 is recommended bnb_4bit_use_double_quant=False, bnb_4bit_quant_type='nf4', ), ) peft_model = PeftModel.from_pretrained( base_model, MODEL_ID, subfolder="loftq_init", is_trainable=True, ) # Do training with peft_model ... ``` ## Experiment Results We have conducted experiments on supervised fine-tuning of [GSM8K](https://huggingface.co/datasets/gsm8k) and [WikiText-2](https://huggingface.co/datasets/wikitext/viewer/wikitext-2-raw-v1). | Model | Bits | Rank | LoRA Initial | GSM8K | WikiText-2 | | -------------- | ---- | ---- | -------------------- | ----- | ---------- | | LLAMA-2-13b | 16 | 64 | Gaussian + 0 | 45.3 | 5.12 | | LLAMA-2-13b | 4 | 64 | Gaussian + 0 (QLoRA) | 39.9 | 5.22 | | **LLAMA-2-13b** | 4 | 64 | LoftQ | 45.0 | 5.16 | **Inference** Here is an example code for inference after the model has been fine-tuned on [GSM8K](https://huggingface.co/datasets/gsm8k). ```python import torch from transformers import AutoModelForCausalLM, BitsAndBytesConfig from peft import PeftModel MODEL_ID = "LoftQ/Llama-2-13b-hf-4bit-64rank" base_model = AutoModelForCausalLM.from_pretrained( MODEL_ID, torch_dtype=torch.bfloat16, # you may change it with different models quantization_config=BitsAndBytesConfig( load_in_4bit=True, bnb_4bit_compute_dtype=torch.bfloat16, # bfloat16 is recommended bnb_4bit_use_double_quant=False, bnb_4bit_quant_type='nf4', ), ) peft_model = PeftModel.from_pretrained( base_model, MODEL_ID, subfolder="gsm8k", is_trainable=True, ) # Do inference with peft_model ... ``` See the full code at our [Github Repo]((https://github.com/yxli2123/LoftQ)) ## Citation ```bibtex @article{li2023loftq, title={Loftq: Lora-fine-tuning-aware quantization for large language models}, author={Li, Yixiao and Yu, Yifan and Liang, Chen and He, Pengcheng and Karampatziakis, Nikos and Chen, Weizhu and Zhao, Tuo}, journal={arXiv preprint arXiv:2310.08659}, year={2023} } ```
{"language": ["en"], "license": "mit", "tags": ["quantization ", "lora"], "pipeline_tag": "text-generation"}
text-generation
ChuGyouk/Llama-2-13b-hf-4bit-64rank
[ "transformers", "safetensors", "llama", "text-generation", "quantization ", "lora", "en", "arxiv:2310.08659", "license:mit", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T15:36:08+00:00
[ "2310.08659" ]
[ "en" ]
TAGS #transformers #safetensors #llama #text-generation #quantization #lora #en #arxiv-2310.08659 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
LoftQ Initialization ==================== | Paper | Code | PEFT Example | LoftQ (LoRA-fine-tuning-aware Quantization) provides a quantized backbone Q and LoRA adapters A and B, given a full-precision pre-trained weight W. This model, 'Llama-2-13b-hf-4bit-64rank', is obtained from LLAMA-2-13b. The backbone is under 'LoftQ/Llama-2-13b-hf-4bit-64rank' and LoRA adapters are under the 'subfolder='loftq\_init''. Model Info ---------- ### Backbone * Stored format: 'torch.bfloat16' * Size: ~ 26 GiB * Loaded format: bitsandbytes nf4 * Size loaded on GPU: ~6.5 GiB ### LoRA adapters * rank: 64 * lora\_alpha: 64 * target\_modules: ["down\_proj", "up\_proj", "q\_proj", "k\_proj", "v\_proj", "o\_proj", "gate\_proj"] Usage ----- Training Here's an example of loading this model and preparing for the LoRA fine-tuning. Experiment Results ------------------ We have conducted experiments on supervised fine-tuning of GSM8K and WikiText-2. Inference Here is an example code for inference after the model has been fine-tuned on GSM8K. See the full code at our Github Repo)
[ "### Backbone\n\n\n* Stored format: 'torch.bfloat16'\n* Size: ~ 26 GiB\n* Loaded format: bitsandbytes nf4\n* Size loaded on GPU: ~6.5 GiB", "### LoRA adapters\n\n\n* rank: 64\n* lora\\_alpha: 64\n* target\\_modules: [\"down\\_proj\", \"up\\_proj\", \"q\\_proj\", \"k\\_proj\", \"v\\_proj\", \"o\\_proj\", \"gate\\_proj\"]\n\n\nUsage\n-----\n\n\nTraining Here's an example of loading this model and preparing for the LoRA fine-tuning.\n\n\nExperiment Results\n------------------\n\n\nWe have conducted experiments on supervised fine-tuning of GSM8K\nand WikiText-2.\n\n\n\nInference Here is an example code for inference after the model has been fine-tuned on GSM8K.\n\n\nSee the full code at our Github Repo)" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #quantization #lora #en #arxiv-2310.08659 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Backbone\n\n\n* Stored format: 'torch.bfloat16'\n* Size: ~ 26 GiB\n* Loaded format: bitsandbytes nf4\n* Size loaded on GPU: ~6.5 GiB", "### LoRA adapters\n\n\n* rank: 64\n* lora\\_alpha: 64\n* target\\_modules: [\"down\\_proj\", \"up\\_proj\", \"q\\_proj\", \"k\\_proj\", \"v\\_proj\", \"o\\_proj\", \"gate\\_proj\"]\n\n\nUsage\n-----\n\n\nTraining Here's an example of loading this model and preparing for the LoRA fine-tuning.\n\n\nExperiment Results\n------------------\n\n\nWe have conducted experiments on supervised fine-tuning of GSM8K\nand WikiText-2.\n\n\n\nInference Here is an example code for inference after the model has been fine-tuned on GSM8K.\n\n\nSee the full code at our Github Repo)" ]
[ 69, 49, 168 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #quantization #lora #en #arxiv-2310.08659 #license-mit #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Backbone\n\n\n* Stored format: 'torch.bfloat16'\n* Size: ~ 26 GiB\n* Loaded format: bitsandbytes nf4\n* Size loaded on GPU: ~6.5 GiB### LoRA adapters\n\n\n* rank: 64\n* lora\\_alpha: 64\n* target\\_modules: [\"down\\_proj\", \"up\\_proj\", \"q\\_proj\", \"k\\_proj\", \"v\\_proj\", \"o\\_proj\", \"gate\\_proj\"]\n\n\nUsage\n-----\n\n\nTraining Here's an example of loading this model and preparing for the LoRA fine-tuning.\n\n\nExperiment Results\n------------------\n\n\nWe have conducted experiments on supervised fine-tuning of GSM8K\nand WikiText-2.\n\n\n\nInference Here is an example code for inference after the model has been fine-tuned on GSM8K.\n\n\nSee the full code at our Github Repo)" ]
[ -0.019112544134259224, 0.048921991139650345, -0.001075908774510026, 0.037833552807569504, 0.04288939759135246, -0.013777052983641624, 0.08079851418733597, 0.14879444241523743, 0.03225483372807503, 0.1245938241481781, 0.09796614199876785, 0.11458219587802887, 0.050111979246139526, 0.18647468090057373, -0.020758669823408127, -0.16290271282196045, 0.02046961337327957, -0.07940264791250229, -0.0005216198624111712, 0.12643511593341827, 0.1102587953209877, -0.047907497733831406, 0.10199088603258133, -0.004186970181763172, -0.003805180313065648, -0.02941203862428665, -0.013793888501822948, -0.0667433887720108, 0.10983481258153915, 0.048318248242139816, 0.006539961788803339, 0.03786640986800194, -0.021621618419885635, -0.22800593078136444, 0.01917674019932747, 0.05799540877342224, 0.046569064259529114, 0.06696971505880356, 0.06131378188729286, 0.02408425696194172, 0.1747795045375824, -0.08809496462345123, -0.035330548882484436, 0.06457141786813736, -0.11261964589357376, -0.17397822439670563, -0.14741627871990204, -0.02646353840827942, 0.08407560735940933, 0.053554631769657135, -0.0517391711473465, 0.09781889617443085, -0.038656920194625854, 0.06510753929615021, 0.20446448028087616, -0.30810320377349854, -0.027813240885734558, 0.14242175221443176, 0.07043883949518204, 0.04631665349006653, -0.05332620069384575, 0.03455934673547745, 0.07457512617111206, -0.0349726639688015, -0.04969716817140579, -0.03135400265455246, -0.03515889495611191, 0.028054798021912575, -0.16418100893497467, -0.05537690967321396, 0.1779910922050476, 0.04812445119023323, -0.0910126343369484, -0.1585480272769928, -0.05358463525772095, -0.15260381996631622, 0.012364618480205536, -0.02633742243051529, -0.0039049589540809393, 0.012529748491942883, -0.0208232831209898, -0.06631356477737427, -0.06729547679424286, -0.11575452983379364, -0.01993217132985592, 0.19593441486358643, 0.05868403613567352, 0.01740325801074505, -0.031044086441397667, 0.1624702662229538, -0.15598537027835846, -0.1080423891544342, -0.12512736022472382, -0.04981764778494835, -0.022946331650018692, 0.0022593385074287653, -0.004302702844142914, 0.005090608727186918, 0.15790444612503052, 0.2934980094432831, -0.042963575571775436, 0.03328181803226471, 0.02644769288599491, 0.020648861303925514, 0.0009915254777297378, 0.11923518776893616, -0.07761044055223465, -0.05809520557522774, 0.06712818890810013, 0.03790442273020744, 0.03411554545164108, -0.02102026343345642, -0.019078334793448448, -0.03749901056289673, 0.014840290881693363, 0.08725486695766449, 0.045966193079948425, 0.06157011166214943, -0.04592437297105789, -0.0002377810305915773, 0.09064551442861557, -0.10694528371095657, -0.049077630043029785, -0.023826925083994865, -0.038133781403303146, -0.098967544734478, 0.01098013948649168, 0.03680923953652382, 0.018651794642210007, 0.01790144480764866, -0.05940823256969452, -0.06695184111595154, -0.0708763524889946, -0.05386381223797798, -0.004459688905626535, -0.0021074931137263775, -0.002720403950661421, -0.1126481145620346, -0.1340956836938858, -0.06460899114608765, -0.013949206098914146, -0.0642266497015953, 0.008701683953404427, 0.007533740717917681, -0.06227404996752739, 0.0022045085206627846, -0.03922177851200104, 0.07452721893787384, -0.07941180467605591, 0.10626555979251862, 0.08645609766244888, 0.07242260873317719, -0.012795946560800076, -0.024429168552160263, -0.12379181385040283, 0.09429962188005447, -0.12231463938951492, 0.08397452533245087, -0.13237427175045013, 0.004189142491668463, -0.1779339611530304, -0.04473153129220009, -0.048764124512672424, 0.008917945437133312, 0.08716575801372528, 0.12007283419370651, -0.12929405272006989, -0.02130485139787197, 0.17964696884155273, -0.07991830259561539, -0.10212629288434982, 0.13392019271850586, 0.015884244814515114, 0.05015168711543083, 0.04294560104608536, 0.17670299112796783, 0.04116429388523102, -0.07600893825292587, -0.0982072725892067, -0.026342468336224556, -0.008439326658844948, 0.061562877148389816, 0.12807044386863708, -0.06775379180908203, -0.042834628373384476, 0.03958113491535187, -0.1047876849770546, 0.030246054753661156, -0.059049297124147415, -0.04101501405239105, -0.0379636287689209, -0.03955599665641785, -0.006851105950772762, -0.041292134672403336, -0.020081521943211555, -0.09651993960142136, -0.14061737060546875, -0.21665844321250916, 0.15313564240932465, -0.016817288473248482, 0.006661885883659124, -0.11851932108402252, 0.158900186419487, -0.02564128302037716, 0.015159955248236656, -0.0828094333410263, -0.027103904634714127, 0.03561178594827652, -0.035877738147974014, 0.04178620129823685, -0.04019361734390259, 0.049441102892160416, 0.049860574305057526, -0.01686004363000393, -0.012222717516124249, -0.0156383216381073, -0.02235543727874756, -0.0733221173286438, -0.1024189218878746, -0.04038047045469284, -0.010733249597251415, 0.11653366684913635, -0.11634491384029388, 0.004476980771869421, 0.14146192371845245, 0.08852756023406982, 0.030663926154375076, -0.06928745657205582, 0.043984804302453995, -0.027397871017456055, -0.05525030568242073, -0.06346000730991364, 0.007103275042027235, -0.04365144670009613, -0.0323539637029171, 0.024131281301379204, -0.07241848856210709, -0.03546331450343132, 0.09229722619056702, 0.06051144748926163, -0.05272123962640762, -0.02825988456606865, -0.05694051459431648, 0.02201886475086212, -0.05510197579860687, -0.013087569735944271, 0.11689804494380951, 0.004832555074244738, 0.05630997568368912, -0.05971279740333557, -0.04784947633743286, -0.01508556492626667, -0.06655961275100708, 0.010456999763846397, 0.11224602907896042, 0.0770280584692955, -0.1160660833120346, 0.062455251812934875, 0.08985183387994766, -0.0251277144998312, 0.17153200507164001, -0.027301838621497154, -0.09806134551763535, -0.05024842545390129, 0.10635077953338623, 0.02414565160870552, 0.036389920860528946, -0.08692795038223267, 0.02409486286342144, 0.06766743957996368, 0.008228219114243984, 0.04851483181118965, -0.13845500349998474, -0.00942546408623457, 0.006693839095532894, -0.06937330961227417, 0.06666631996631622, 0.10033494979143143, -0.04448782280087471, 0.07964153587818146, 0.02574225328862667, 0.022228986024856567, 0.002903413027524948, -0.025749942287802696, -0.0498538613319397, 0.18446290493011475, -0.06546098738908768, -0.10842932015657425, -0.15001997351646423, -0.04805072396993637, -0.06693119555711746, 0.009062390774488449, 0.06217841058969498, -0.13770368695259094, -0.0821065604686737, -0.05669139698147774, 0.12942959368228912, 0.07286826521158218, 0.02993641421198845, -0.025936348363757133, 0.027707483619451523, 0.051799334585666656, -0.15525060892105103, -0.016824107617139816, 0.014102976769208908, -0.1460096836090088, 0.06119503080844879, 0.04715236648917198, 0.09712889790534973, 0.10123215615749359, 0.03328682854771614, 0.026736292988061905, 0.008160886354744434, 0.2919631600379944, -0.09365608543157578, 0.008158191107213497, 0.24516341090202332, 0.03369716927409172, -0.007129654288291931, 0.09143515676259995, 0.006801762618124485, -0.13471993803977966, 0.07655291259288788, 0.07250555604696274, -0.0597497932612896, -0.22454337775707245, -0.04327715188264847, -0.03342336788773537, 0.0501197911798954, 0.010330427438020706, 0.08257223665714264, 0.04785420000553131, 0.09904627501964569, -0.056530941277742386, -0.006115288473665714, 0.024837229400873184, 0.07706785947084427, 0.05890896916389465, -0.02002522349357605, 0.08840245753526688, -0.04191632941365242, -0.03877394646406174, 0.1032816618680954, 0.05497157573699951, 0.0941615104675293, -0.03450165316462517, 0.15208692848682404, 0.11752281337976456, 0.12532947957515717, 0.05003767088055611, -0.0006235899636521935, -0.022626202553510666, -0.0076372697949409485, -0.009955808520317078, -0.08588344603776932, -0.0941820740699768, 0.066864013671875, -0.024164041504263878, -0.057065535336732864, 0.013690036721527576, 0.038374848663806915, 0.07929019629955292, 0.10212193429470062, 0.14899587631225586, -0.235630601644516, -0.055752743035554886, 0.04199284687638283, -0.054188381880521774, -0.06862527132034302, 0.0792582556605339, 0.12007394433021545, -0.08445359766483307, 0.020257091149687767, -0.03865789622068405, 0.05860550329089165, -0.13232716917991638, 0.015289238654077053, 0.05972398817539215, 0.09678676724433899, -0.04091509431600571, 0.10666461288928986, -0.21410945057868958, 0.172893688082695, 0.023338817059993744, 0.03243974596261978, -0.06132199615240097, 0.049378957599401474, 0.02705271728336811, -0.04918861761689186, 0.15614087879657745, -0.03505224734544754, -0.040408674627542496, -0.14370115101337433, -0.12815538048744202, 0.038980793207883835, 0.10265696048736572, -0.12964409589767456, 0.13714131712913513, -0.047621194273233414, 0.013544409535825253, -0.039176300168037415, 0.012658433988690376, -0.12828463315963745, -0.09310171008110046, 0.07361115515232086, 0.05018497630953789, 0.09218882024288177, -0.12587496638298035, -0.00904939603060484, -0.0698709711432457, 0.17339260876178741, -0.07552330940961838, -0.07028041034936905, -0.09660376608371735, 0.0730115994811058, 0.14307115972042084, -0.0923687070608139, 0.01569029502570629, -0.0038828132674098015, 0.032798074185848236, -0.0016286963364109397, -0.05772671848535538, 0.09737343341112137, -0.055160459131002426, -0.21358463168144226, -0.004423889331519604, 0.15095758438110352, -0.014790678396821022, 0.021931827068328857, -0.009058481082320213, 0.05784762650728226, -0.01505495235323906, -0.13064727187156677, 0.06249430403113365, 0.0473879911005497, -0.027980277314782143, 0.09261219203472137, 0.04232097044587135, 0.0507890060544014, -0.06404602527618408, -0.07300122827291489, 0.12088722735643387, 0.3499537706375122, -0.029042372480034828, 0.033384330570697784, -0.0009301595855504274, -0.01615241728723049, -0.22368355095386505, -0.041800908744335175, 0.017146944999694824, 0.015552309341728687, -0.0020120053086429834, -0.12807698547840118, 0.06056571006774902, 0.11116304993629456, -0.06604693084955215, 0.13058631122112274, -0.27937909960746765, -0.15858720242977142, 0.09588164836168289, 0.060425251722335815, 0.10432233661413193, -0.16018012166023254, -0.09338679164648056, -0.0016821724129840732, -0.1751452535390854, 0.016893967986106873, -0.1194545105099678, 0.12840873003005981, -0.00965301413089037, 0.008755774237215519, 0.040964268147945404, -0.06537782400846481, 0.20694629848003387, -0.04600312188267708, 0.03304712846875191, -0.07287802547216415, -0.018636533990502357, -0.011900044046342373, -0.03395731747150421, 0.04641032963991165, -0.037094131112098694, 0.057070668786764145, -0.11034891754388809, -0.05424952134490013, -0.03995375707745552, 0.07737066596746445, -0.0008820901857689023, -0.0865129828453064, -0.13682672381401062, 0.07495590299367905, -0.0413702167570591, 0.020455915480852127, 0.12194734811782837, -0.06643781810998917, 0.04202131927013397, 0.1398508995771408, 0.05914577096700668, 0.03394819423556328, -0.025082267820835114, 0.049481794238090515, 0.018344376236200333, 0.05025113373994827, -0.1322736293077469, 0.019882703199982643, 0.0742894783616066, 0.05486306920647621, 0.06611982733011246, 0.025805186480283737, -0.10518451035022736, 0.04905865713953972, 0.09816037118434906, -0.14859530329704285, -0.13975197076797485, -0.06805745512247086, -0.03955816105008125, -0.05459723621606827, 0.05127907544374466, 0.16290871798992157, 0.010606074705719948, -0.021991316229104996, 0.00018285673286300153, 0.09350117295980453, -0.07572103291749954, 0.18010766804218292, 0.04545802250504494, 0.06561083346605301, -0.10233563929796219, 0.02529091387987137, 0.07406903803348541, 0.027513382956385612, 0.03553596884012222, 0.03948143869638443, -0.0701361745595932, -0.03867504373192787, -0.07179377973079681, -0.007742453832179308, -0.020899878814816475, -0.03561415895819664, -0.05522249639034271, -0.0608249306678772, -0.01277577318251133, -0.02670610137283802, 0.08398234844207764, 0.06559492647647858, 0.0302206352353096, -0.005657045636326075, -0.08099709451198578, 0.1149883046746254, -0.0020138362888246775, 0.08781297504901886, -0.16977405548095703, 0.03315045312047005, -0.05050192400813103, 0.007243331987410784, -0.02968676947057247, -0.0015128368977457285, -0.1328156292438507, -0.007428628858178854, -0.09473942220211029, -0.03527185320854187, -0.08387765288352966, -0.023303456604480743, 0.026052216067910194, -0.048295389860868454, -0.03810909762978554, 0.028278183192014694, -0.09205044060945511, -0.045533645898103714, -0.05469302833080292, 0.08889076858758926, -0.11165177077054977, 0.017562998458743095, 0.08234963566064835, -0.08083122968673706, 0.11509867012500763, -0.015144525095820427, 0.04119228199124336, 0.03741038963198662, -0.10719635337591171, 0.024679135531187057, 0.04306711256504059, -0.00007032115536276251, -0.021352745592594147, -0.12964355945587158, 0.023090286180377007, -0.0438610278069973, 0.0015512766549363732, 0.019132105633616447, 0.06860100477933884, -0.13814447820186615, -0.02255415730178356, -0.03740866482257843, 0.01448861975222826, -0.020972179248929024, 0.03236166387796402, 0.056683603674173355, 0.035113636404275894, 0.12212350964546204, -0.07303867489099503, -0.006640324834734201, -0.15572743117809296, 0.01410598959773779, -0.004396103788167238, -0.07057525962591171, -0.049020376056432724, -0.010718226432800293, 0.07513763010501862, -0.048325665295124054, 0.1689828485250473, 0.014002406969666481, -0.0754815861582756, 0.005127924028784037, -0.04019112139940262, 0.14430981874465942, 0.01221314538270235, 0.23332713544368744, 0.05598803609609604, 0.03186718747019768, -0.03169698268175125, 0.013962339609861374, 0.05469699203968048, -0.0895903930068016, 0.1692541241645813, 0.13923217356204987, -0.04239809513092041, 0.03739453852176666, 0.022391049191355705, -0.13244789838790894, -0.04628358408808708, 0.07537857443094254, -0.059180911630392075, 0.06921158730983734, -0.051103994250297546, 0.03436331823468208, 0.1394202709197998, -0.14418044686317444, 0.019242946058511734, 0.0017585407476872206, -0.029254863038659096, -0.14803266525268555, -0.16102972626686096, -0.08509010076522827, -0.17964762449264526, 0.039218075573444366, -0.11161870509386063, 0.052154816687107086, 0.0737946480512619, 0.03304622694849968, -0.032456208020448685, 0.19431228935718536, -0.09427667409181595, -0.04641139879822731, 0.03069411963224411, -0.023584501817822456, -0.029389282688498497, -0.014315496198832989, -0.07213979959487915, 0.04246106743812561, 0.0072104427963495255, 0.06859022378921509, -0.0028385608457028866, -0.0558747872710228, 0.02368904836475849, -0.030902408063411713, -0.050970107316970825, -0.04546838626265526, 0.05367922782897949, 0.04878294840455055, 0.09455213695764542, 0.03086077608168125, -0.040335942059755325, 0.013374646194279194, 0.24282018840312958, -0.04314348101615906, -0.07090726494789124, -0.11951790750026703, 0.1465451717376709, -0.02076154574751854, 0.03884126618504524, 0.0016882203053683043, -0.12066703289747238, 0.013193962164223194, 0.19781909883022308, 0.1444985717535019, 0.010559565387666225, -0.01631903275847435, 0.02583617903292179, 0.009581481106579304, 0.006584026385098696, 0.06945528835058212, 0.07410835474729538, 0.17256447672843933, -0.04259215667843819, -0.04478694126009941, 0.04069254547357559, -0.03006717935204506, -0.024610325694084167, 0.0852280706167221, 0.009935298003256321, -0.05712869390845299, -0.041595276445150375, 0.06275290995836258, -0.07140184193849564, -0.12282326072454453, 0.0037701588589698076, -0.11155589669942856, -0.13912326097488403, -0.02842577174305916, 0.03346330299973488, 0.0010736386757344007, 0.04705424606800079, -0.01603083498775959, -0.03057490475475788, 0.03837992995977402, 0.004286128096282482, -0.10853640735149384, -0.045582085847854614, 0.06699150800704956, 0.01105460524559021, 0.22340020537376404, -0.005284398328512907, 0.06820344924926758, 0.13383616507053375, -0.07062867283821106, -0.14523214101791382, 0.13211771845817566, 0.038555044680833817, -0.13395138084888458, 0.04249817878007889, 0.11240540444850922, 0.026066996157169342, 0.08267275243997574, 0.07188045233488083, -0.1552945226430893, 0.014231779612600803, 0.08581278473138809, 0.00405194703489542, -0.1686766892671585, 0.02312624454498291, -0.1561184674501419, 0.1589280068874359, 0.15014337003231049, -0.0808420181274414, 0.045395322144031525, -0.056725043803453445, 0.0993976891040802, 0.0029196112882345915, 0.09861448407173157, -0.006847381591796875, -0.1706659197807312, -0.00441908510401845, -0.03563644364476204, -0.017603090032935143, -0.21832555532455444, -0.0218844972550869, -0.01834007538855076, 0.024420466274023056, -0.05727565288543701, 0.11888976395130157, 0.1813814491033554, 0.013355720788240433, -0.021298889070749283, -0.11318770051002502, -0.023860735818743706, 0.11282533407211304, -0.1575356423854828, -0.06581888347864151 ]
null
null
stable-baselines3
# **PPO** Agent playing **LunarLander-v2** This is a trained model of a **PPO** agent playing **LunarLander-v2** using the [stable-baselines3 library](https://github.com/DLR-RM/stable-baselines3). ## Usage (with Stable-baselines3) TODO: Add your code ```python from stable_baselines3 import ... from huggingface_sb3 import load_from_hub ... ```
{"library_name": "stable-baselines3", "tags": ["LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "stable-baselines3"], "model-index": [{"name": "PPO", "results": [{"task": {"type": "reinforcement-learning", "name": "reinforcement-learning"}, "dataset": {"name": "LunarLander-v2", "type": "LunarLander-v2"}, "metrics": [{"type": "mean_reward", "value": "261.49 +/- 16.32", "name": "mean_reward", "verified": false}]}]}]}
reinforcement-learning
michaelcheung/ppo-LunarLander-v2
[ "stable-baselines3", "LunarLander-v2", "deep-reinforcement-learning", "reinforcement-learning", "model-index", "region:us" ]
2024-02-10T15:51:40+00:00
[]
[]
TAGS #stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us
# PPO Agent playing LunarLander-v2 This is a trained model of a PPO agent playing LunarLander-v2 using the stable-baselines3 library. ## Usage (with Stable-baselines3) TODO: Add your code
[ "# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.", "## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ "TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n", "# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.", "## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ 39, 41, 17 ]
[ "passage: TAGS\n#stable-baselines3 #LunarLander-v2 #deep-reinforcement-learning #reinforcement-learning #model-index #region-us \n# PPO Agent playing LunarLander-v2\nThis is a trained model of a PPO agent playing LunarLander-v2\nusing the stable-baselines3 library.## Usage (with Stable-baselines3)\nTODO: Add your code" ]
[ 0.03942384943366051, 0.04900386184453964, -0.005304091144353151, 0.026427261531352997, 0.107408307492733, -0.026511888951063156, 0.11188238859176636, 0.0814051404595375, 0.10722193866968155, 0.04762078449130058, 0.08338645845651627, 0.06030960753560066, 0.05080918222665787, 0.2571701407432556, 0.04754156619310379, -0.22987541556358337, 0.036159250885248184, -0.04869936779141426, 0.12395193427801132, 0.07178173214197159, -0.0038484656251966953, -0.06485428661108017, 0.020415637642145157, -0.013290755450725555, 0.05367108806967735, 0.04282612353563309, -0.01716216839849949, -0.08207534998655319, 0.07169748842716217, -0.06345846503973007, 0.06986866891384125, 0.07677983492612839, 0.13218913972377777, -0.17832116782665253, 0.029566360637545586, 0.02571309357881546, -0.07189024239778519, 0.01342033501714468, 0.008019951172173023, 0.05120139941573143, 0.17303818464279175, 0.019879888743162155, 0.07844575494527817, -0.0025605305563658476, -0.15412317216396332, -0.018950799480080605, 0.0436202734708786, 0.12546207010746002, 0.08808347582817078, 0.04605821147561073, 0.01970590092241764, 0.17503218352794647, -0.054352790117263794, -0.028833400458097458, 0.21759237349033356, -0.2881564497947693, -0.031460098922252655, 0.321048766374588, 0.06997483223676682, 0.09725230932235718, -0.07540661096572876, -0.03619609400629997, 0.007783263456076384, -0.013137873262166977, -0.028666524216532707, -0.07447073608636856, 0.17313385009765625, 0.05152064561843872, -0.05057951435446739, -0.09541505575180054, 0.16948209702968597, 0.006921638268977404, 0.0018855923553928733, -0.019282981753349304, 0.009060598909854889, 0.07402525842189789, -0.016097044572234154, -0.07255112379789352, 0.057438433170318604, 0.05330665782094002, 0.019649166613817215, -0.1435653269290924, -0.10762494057416916, -0.022740179672837257, -0.008012006990611553, 0.17786912620067596, -0.009255532175302505, 0.042902372777462006, 0.003065188182517886, 0.10384012013673782, -0.12480384111404419, -0.03354184702038765, -0.0454259067773819, -0.07565800100564957, -0.0223417766392231, -0.02058211714029312, -0.03580251708626747, 0.07184842973947525, 0.11971849203109741, 0.027368178591132164, 0.09350208193063736, 0.047715865075588226, -0.03206788748502731, 0.06343851238489151, 0.05555703118443489, 0.14222665131092072, 0.05807621404528618, 0.012854371219873428, 0.13179877400398254, 0.055213116109371185, 0.033023182302713394, -0.0613492950797081, -0.18252409994602203, 0.07489913702011108, -0.07031869143247604, 0.007941240444779396, 0.12051256000995636, -0.04480670019984245, -0.1183447614312172, -0.037500523030757904, -0.017392054200172424, -0.06224250793457031, -0.025395862758159637, 0.0547584593296051, -0.02883218228816986, -0.03973718360066414, 0.0011496668448671699, 0.09384800493717194, 0.00953749567270279, -0.1752052903175354, 0.03303423151373863, -0.025042934343218803, -0.10782608389854431, 0.009975161403417587, 0.0022444494534283876, 0.03394931182265282, 0.04408763721585274, -0.11822668462991714, -0.30899152159690857, -0.07652641832828522, 0.05490870401263237, -0.06516939401626587, -0.18425025045871735, -0.13193942606449127, 0.02454492449760437, -0.09037084132432938, -0.044885024428367615, -0.12759265303611755, -0.028549788519740105, 0.01743689924478531, 0.011519349180161953, 0.10758619755506516, -0.0106219332665205, -0.012188062071800232, -0.1571401208639145, 0.008273907005786896, -0.20951123535633087, 0.0890483483672142, -0.019150104373693466, 0.037884220480918884, -0.032381169497966766, -0.07404014468193054, 0.030707746744155884, 0.052499737590551376, -0.01474119070917368, 0.13510210812091827, -0.15592676401138306, -0.03691192343831062, -0.007996266707777977, -0.13611900806427002, -0.04786273464560509, -0.10358831286430359, -0.04357128217816353, 0.13354332745075226, 0.018664736300706863, 0.15356586873531342, -0.08709818124771118, -0.0722038671374321, 0.20489206910133362, -0.010411538183689117, -0.12820468842983246, -0.076752208173275, 0.10165707021951675, 0.021510310471057892, -0.056606587022542953, -0.02523270808160305, -0.1839766949415207, -0.0152357779443264, -0.04550420492887497, -0.047039128839969635, 0.01796751655638218, -0.010888241231441498, 0.13837894797325134, 0.08494598418474197, 0.05018039792776108, -0.06086122244596481, -0.006730288732796907, 0.10779471695423126, 0.08823856711387634, 0.008680110797286034, 0.023406028747558594, -0.05774238705635071, 0.09552932530641556, -0.04003755748271942, -0.0142367510125041, -0.08283266425132751, -0.036246106028556824, -0.026256313547492027, 0.17507147789001465, 0.09440762549638748, 0.2257927656173706, 0.09567736834287643, 0.039160262793302536, 0.031270865350961685, -0.13181598484516144, -0.1425403207540512, -0.0017254541162401438, 0.09020978957414627, -0.14270411431789398, -0.04119925573468208, -0.08974775671958923, -0.17768175899982452, -0.12202505767345428, 0.0006432619411498308, -0.17960017919540405, 0.06390921026468277, 0.05408334732055664, -0.035177867859601974, 0.03272094577550888, 0.13032332062721252, -0.011533179320394993, -0.03967514634132385, 0.0831870287656784, 0.0379033200442791, -0.041234664618968964, -0.021742934361100197, 0.11885567009449005, 0.15673065185546875, 0.13124459981918335, -0.03511447086930275, 0.004914294462651014, 0.07076404243707657, -0.02309088408946991, 0.06539414077997208, 0.0558244064450264, 0.20973342657089233, 0.188301220536232, 0.038996949791908264, 0.008822928182780743, -0.07048165798187256, 0.0855446457862854, -0.0742373839020729, -0.14302679896354675, -0.05579735338687897, 0.08729292452335358, 0.016605578362941742, 0.023469142615795135, 0.08711627870798111, 0.024545932188630104, 0.09132762253284454, 0.15968108177185059, 0.01990218088030815, -0.09659269452095032, -0.050218869000673294, 0.01175848301500082, 0.027713103219866753, 0.04794301092624664, -0.04514073207974434, -0.00937939714640379, 0.017020760104060173, -0.10303554683923721, 0.031789086759090424, -0.1413339376449585, -0.1358717679977417, 0.044326696544885635, 0.003906996920704842, 0.010907664895057678, 0.02786896750330925, -0.0038291432429105043, 0.019039705395698547, 0.04351753741502762, -0.06975466758012772, 0.047416772693395615, -0.024745507165789604, -0.020031947642564774, 0.03340689837932587, -0.057257164269685745, -0.205775648355484, -0.17696654796600342, 0.00013708483311347663, -0.09910997003316879, 0.10194740444421768, 0.018308809027075768, -0.12373185902833939, 0.047737859189510345, -0.05822649225592613, 0.027574289590120316, -0.01875593699514866, -0.049130141735076904, 0.10507171601057053, 0.1525275856256485, -0.016146350651979446, 0.018018173053860664, -0.04865182936191559, -0.10157987475395203, -0.19632206857204437, 0.0691583976149559, 0.04680244252085686, 0.014610917307436466, 0.10669491440057755, 0.018072687089443207, 0.02367905154824257, -0.007674071006476879, -0.016521066427230835, -0.011659215204417706, -0.08781040459871292, 0.31909599900245667, 0.04510033503174782, -0.025173069909214973, 0.02041010931134224, -0.0043001663871109486, -0.028083480894565582, 0.03263787180185318, -0.0985708013176918, -0.07548979669809341, -0.08774089068174362, -0.04367410019040108, -0.09784720093011856, 0.053299110382795334, 0.05916472524404526, 0.003188040340319276, -0.07727594673633575, 0.04221395403146744, 0.11369874328374863, -0.0923808291554451, -0.07137343287467957, 0.07477962225675583, 0.0972946360707283, -0.07331304252147675, 0.00012658814375754446, 0.00874367356300354, 0.023951783776283264, 0.037102166563272476, 0.06778035312891006, -0.03966575115919113, 0.08589404821395874, -0.19917890429496765, 0.0372927263379097, 0.106058269739151, 0.023754918947815895, 0.0638108178973198, 0.07643651217222214, -0.1058402881026268, -0.008500572293996811, -0.032518330961465836, -0.21341575682163239, 0.1668180525302887, 0.1355515867471695, 0.06788124144077301, -0.025637222453951836, -0.00461410591378808, -0.0649740919470787, 0.05773647129535675, 0.02723747305572033, -0.14758841693401337, 0.004883295856416225, 0.06064270809292793, 0.026899009943008423, 0.01614922471344471, 0.07971042394638062, 0.014697225764393806, -0.1801026314496994, -0.014406266622245312, 0.10730406641960144, 0.002390873385593295, 0.0053148469887673855, -0.03175045922398567, -0.1755964607000351, 0.0751047357916832, 0.004285442177206278, 0.07233936339616776, -0.1676585078239441, 0.14297930896282196, -0.10089799761772156, 0.07726949453353882, -0.004285062663257122, -0.021311495453119278, 0.02507244050502777, -0.0541163794696331, 0.15163759887218475, 0.01058570109307766, -0.021810131147503853, -0.1200498715043068, -0.1717042326927185, -0.019227758049964905, -0.11788936704397202, -0.11679866164922714, 0.050424277782440186, 0.062185097485780716, 0.04923136904835701, -0.061147067695856094, 0.1518532931804657, -0.047422297298908234, 0.060713399201631546, -0.06893875449895859, -0.06755045056343079, 0.03764858841896057, -0.12588608264923096, -0.08176055550575256, 0.05573027580976486, 0.19166934490203857, 0.15833087265491486, -0.02816431224346161, -0.03472423925995827, -0.047419581562280655, -0.006212298292666674, -0.007802055217325687, 0.0275666993111372, 0.023223137483000755, 0.07315318286418915, -0.07681374251842499, -0.11649256944656372, 0.033787861466407776, -0.06713802367448807, -0.055589709430933, -0.015439179725944996, 0.1513158082962036, 0.04671623185276985, 0.07720734924077988, -0.018946662545204163, 0.03887668624520302, -0.001724981120787561, -0.056474871933460236, 0.16197094321250916, 0.03885216265916824, -0.05193585529923439, 0.06837689876556396, 0.053174007683992386, 0.043745119124650955, 0.03011113777756691, -0.026783017441630363, 0.206032395362854, 0.1980147808790207, 0.014206883497536182, 0.2175983190536499, 0.03177616000175476, -0.03772832080721855, -0.1300560086965561, -0.065880686044693, -0.006372632458806038, 0.03559038043022156, 0.08070417493581772, -0.18207235634326935, -0.015011128038167953, -0.05689644813537598, -0.034518610686063766, -0.15059494972229004, -0.28553900122642517, -0.05957856774330139, 0.20075850188732147, 0.14706264436244965, 0.27519428730010986, -0.10432573407888412, 0.035197313874959946, 0.02663275972008705, -0.04912831634283066, -0.006501141935586929, 0.00018665487004909664, 0.10268618166446686, -0.15421873331069946, 0.1176437959074974, 0.08486983180046082, -0.019002694636583328, 0.01058861706405878, -0.1619086116552353, 0.00936629343777895, -0.12191236019134521, 0.05354422330856323, 0.1400289237499237, -0.048128653317689896, -0.054873593151569366, 0.14033560454845428, -0.024562934413552284, -0.22685599327087402, -0.04648222774267197, -0.043600670993328094, -0.010640020482242107, 0.026607351377606392, -0.1013401448726654, 0.04101909324526787, 0.1330099105834961, 0.009380043484270573, 0.1147187277674675, 0.11749245226383209, -0.052566803991794586, 0.10792597383260727, 0.2257719188928604, -0.018785694614052773, 0.04689010605216026, -0.12743118405342102, -0.0012336712097749114, -0.028270328417420387, 0.013657891191542149, -0.09504974633455276, -0.09938385337591171, 0.02366873063147068, 0.02872389927506447, 0.009118586778640747, 0.0921793207526207, -0.029922157526016235, 0.0759170651435852, 0.06817561388015747, -0.13014446198940277, -0.16288450360298157, 0.015828335657715797, -0.007344507612287998, 0.08354310691356659, 0.00027861111448146403, 0.08878035843372345, -0.11932205408811569, -0.018093237653374672, -0.03153328225016594, -0.03319635987281799, -0.130486860871315, -0.07138993591070175, 0.06156524643301964, 0.028095467016100883, -0.06602972000837326, 0.1398407518863678, 0.026440169662237167, 0.15942534804344177, 0.049197953194379807, 0.012499804608523846, 0.07227300107479095, -0.05345509201288223, 0.1283530443906784, 0.13818155229091644, -0.00868943240493536, -0.05460423603653908, -0.1013643890619278, -0.10236792266368866, 0.08925779908895493, -0.05773641914129257, 0.07476430386304855, -0.14885357022285461, -0.06675903499126434, 0.015772046521306038, 0.016141414642333984, -0.09562095999717712, 0.02571965754032135, -0.01625603251159191, -0.18119946122169495, 0.056570518761873245, -0.048285093158483505, 0.0440407395362854, -0.06347788125276566, -0.1110161691904068, -0.17226378619670868, 0.06091433763504028, 0.08593481779098511, -0.053876690566539764, -0.12229149043560028, 0.011023230850696564, -0.00012518465518951416, -0.06341652572154999, -0.05023367330431938, 0.09722746908664703, -0.11020902544260025, 0.031452205032110214, -0.012567701749503613, 0.08853451162576675, -0.03510405123233795, -0.011538895778357983, 0.044220831245183945, -0.08039166033267975, -0.009481523185968399, 0.03534642979502678, -0.026372017338871956, -0.04127239063382149, -0.2689029574394226, 0.0036654395516961813, 0.0341104120016098, 0.02497158572077751, 0.07856601476669312, 0.011906822212040424, 0.021174922585487366, 0.03993808850646019, -0.15396519005298615, -0.013395369984209538, 0.14574195444583893, -0.07689505815505981, -0.022186370566487312, 0.05703273415565491, -0.09054436534643173, 0.013882770203053951, -0.030287226662039757, 0.1345842480659485, 0.023923413828015327, 0.06404478847980499, -0.0851147472858429, 0.10106813907623291, -0.1451139897108078, -0.04998219385743141, -0.01244612317532301, 0.09761348366737366, 0.07019034773111343, -0.10272270441055298, 0.014697125181555748, 0.04210108891129494, 0.19416837394237518, 0.016384804621338844, -0.0356343574821949, -0.03396720811724663, 0.004015897400677204, 0.22076453268527985, 0.03044266067445278, 0.10457023978233337, 0.07281364500522614, -0.026583973318338394, 0.12624378502368927, 0.09929762035608292, 0.11280370503664017, -0.055645186454057693, 0.13904185593128204, 0.04667386785149574, 0.038641396909952164, 0.0614289753139019, 0.06836545467376709, 0.09098632633686066, -0.0008288522367365658, 0.1138714924454689, 0.013811973854899406, -0.02422109805047512, -0.021335409954190254, 0.17759373784065247, 0.10501719266176224, -0.14769648015499115, 0.029047364369034767, -0.01258957851678133, 0.039933037012815475, -0.014194529503583908, -0.15634691715240479, -0.07240267097949982, -0.3315149247646332, 0.1226184144616127, -0.07119352370500565, 0.019930170848965645, 0.007913772016763687, -0.037425633519887924, -0.03296699747443199, -0.04477746784687042, 0.13151589035987854, -0.013641550205647945, -0.006079165264964104, -0.04815853759646416, -0.015360191464424133, -0.11607866734266281, -0.11200575530529022, -0.013207737356424332, -0.13671602308750153, -0.010119039565324783, 0.05595948174595833, 0.003977729007601738, 0.01821410097181797, -0.03142618387937546, 0.0024383175186812878, 0.06541839241981506, -0.05751744285225868, 0.056182678788900375, 0.12097269296646118, 0.08766137808561325, -0.1058853268623352, 0.031048951670527458, 0.2011747509241104, 0.04359564557671547, -0.12483977526426315, 0.01449228823184967, 0.1819491684436798, 0.004885740112513304, 0.017068125307559967, -0.006097703706473112, -0.0540788508951664, -0.07554277032613754, 0.1251034289598465, 0.08296554535627365, -0.09985227137804031, 0.015833314508199692, -0.0726347416639328, -0.01594804972410202, -0.06374675035476685, 0.10130585730075836, 0.09538925439119339, 0.04440245032310486, -0.10621760785579681, -0.08487539738416672, -0.10891728103160858, 0.040588874369859695, -0.08629853278398514, -0.07311757653951645, 0.09629398584365845, -0.07057105004787445, -0.07029950618743896, 0.025521177798509598, -0.17978744208812714, -0.009467960335314274, 0.1711762249469757, -0.24654000997543335, -0.0916430801153183, -0.10857923328876495, 0.14477859437465668, 0.016497576609253883, 0.1013975441455841, -0.006207061931490898, -0.007889035157859325, -0.20577777922153473, 0.024890204891562462, -0.05293011665344238, -0.02073732763528824, 0.07814782857894897, -0.09476397186517715, 0.22629831731319427, -0.08276885002851486, 0.020940175279974937, 0.012659613974392414, 0.0870661810040474, -0.030675338581204414, 0.09283176809549332, -0.03660329803824425, -0.12576518952846527, -0.03620953485369682, 0.03001813031733036, 0.013904244638979435, 0.10071761906147003, 0.09772487729787827, -0.03414725139737129, 0.03389119729399681, 0.09747414290904999, 0.04172342270612717, -0.023843804374337196, 0.0360250361263752, -0.17077107727527618, 0.02182629331946373, -0.018498148769140244, -0.06935930997133255, 0.03687669709324837, -0.06603235751390457, 0.1639697551727295, 0.04022442549467087, 0.0670473501086235, -0.036152735352516174, 0.0073931049555540085, -0.014454689808189869, -0.013775371946394444, -0.026180334389209747, -0.17259705066680908, -0.10422050207853317, -0.1347656100988388, -0.012701659463346004, -0.034971047192811966, 0.04591470584273338, 0.023234914988279343, -0.0003200018545612693, -0.014577031135559082, -0.12090865522623062, 0.04360328987240791, 0.11146783083677292, -0.04631396010518074, -0.026193076744675636 ]
null
null
transformers
# capybagel-7B-slerp capybagel-7B-slerp is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing): * [jondurbin/bagel-dpo-7b-v0.4](https://huggingface.co/jondurbin/bagel-dpo-7b-v0.4) * [argilla/CapybaraHermes-2.5-Mistral-7B](https://huggingface.co/argilla/CapybaraHermes-2.5-Mistral-7B) ## 🧩 Configuration ```yaml slices: - sources: - model: jondurbin/bagel-dpo-7b-v0.4 layer_range: [0, 32] - model: argilla/CapybaraHermes-2.5-Mistral-7B layer_range: [0, 32] merge_method: slerp base_model: jondurbin/bagel-dpo-7b-v0.4 parameters: t: - filter: self_attn value: [0, 0.5, 0.3, 0.7, 1] - filter: mlp value: [1, 0.5, 0.7, 0.3, 0] - value: 0.5 dtype: bfloat16 ``` ## 💻 Usage ```python !pip install -qU transformers accelerate from transformers import AutoTokenizer import transformers import torch model = "jacobi/capybagel-7B-slerp" messages = [{"role": "user", "content": "What is a large language model?"}] tokenizer = AutoTokenizer.from_pretrained(model) prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) pipeline = transformers.pipeline( "text-generation", model=model, torch_dtype=torch.float16, device_map="auto", ) outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95) print(outputs[0]["generated_text"]) ```
{"tags": ["merge", "mergekit", "lazymergekit", "jondurbin/bagel-dpo-7b-v0.4", "argilla/CapybaraHermes-2.5-Mistral-7B"], "base_model": ["jondurbin/bagel-dpo-7b-v0.4", "argilla/CapybaraHermes-2.5-Mistral-7B"]}
text-generation
jacobi/capybagel-7B-slerp
[ "transformers", "safetensors", "mistral", "text-generation", "merge", "mergekit", "lazymergekit", "jondurbin/bagel-dpo-7b-v0.4", "argilla/CapybaraHermes-2.5-Mistral-7B", "conversational", "base_model:jondurbin/bagel-dpo-7b-v0.4", "base_model:argilla/CapybaraHermes-2.5-Mistral-7B", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T15:53:07+00:00
[]
[]
TAGS #transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #jondurbin/bagel-dpo-7b-v0.4 #argilla/CapybaraHermes-2.5-Mistral-7B #conversational #base_model-jondurbin/bagel-dpo-7b-v0.4 #base_model-argilla/CapybaraHermes-2.5-Mistral-7B #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# capybagel-7B-slerp capybagel-7B-slerp is a merge of the following models using LazyMergekit: * jondurbin/bagel-dpo-7b-v0.4 * argilla/CapybaraHermes-2.5-Mistral-7B ## Configuration ## Usage
[ "# capybagel-7B-slerp\n\ncapybagel-7B-slerp is a merge of the following models using LazyMergekit:\n* jondurbin/bagel-dpo-7b-v0.4\n* argilla/CapybaraHermes-2.5-Mistral-7B", "## Configuration", "## Usage" ]
[ "TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #jondurbin/bagel-dpo-7b-v0.4 #argilla/CapybaraHermes-2.5-Mistral-7B #conversational #base_model-jondurbin/bagel-dpo-7b-v0.4 #base_model-argilla/CapybaraHermes-2.5-Mistral-7B #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# capybagel-7B-slerp\n\ncapybagel-7B-slerp is a merge of the following models using LazyMergekit:\n* jondurbin/bagel-dpo-7b-v0.4\n* argilla/CapybaraHermes-2.5-Mistral-7B", "## Configuration", "## Usage" ]
[ 134, 62, 4, 3 ]
[ "passage: TAGS\n#transformers #safetensors #mistral #text-generation #merge #mergekit #lazymergekit #jondurbin/bagel-dpo-7b-v0.4 #argilla/CapybaraHermes-2.5-Mistral-7B #conversational #base_model-jondurbin/bagel-dpo-7b-v0.4 #base_model-argilla/CapybaraHermes-2.5-Mistral-7B #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# capybagel-7B-slerp\n\ncapybagel-7B-slerp is a merge of the following models using LazyMergekit:\n* jondurbin/bagel-dpo-7b-v0.4\n* argilla/CapybaraHermes-2.5-Mistral-7B## Configuration## Usage" ]
[ -0.07912690192461014, 0.04069589450955391, -0.005559020210057497, 0.0073607973754405975, 0.043064143508672714, 0.0411258302628994, 0.20931696891784668, 0.07104162126779556, 0.0983128622174263, 0.056783802807331085, 0.08934354782104492, 0.14201298356056213, 0.03674941137433052, 0.13416600227355957, -0.0430685356259346, -0.18913567066192627, 0.06579795479774475, 0.023847509175539017, -0.042808856815099716, 0.10548466444015503, 0.11017845571041107, -0.03956587612628937, 0.1077854111790657, 0.006445900071412325, -0.02422061562538147, -0.00847594439983368, 0.005527201108634472, -0.0612245574593544, 0.12087946385145187, 0.05571984127163887, 0.12365052103996277, 0.08992007374763489, -0.021096015349030495, -0.13037534058094025, 0.03726980835199356, -0.014606140553951263, -0.01991337165236473, 0.09056832641363144, 0.0565091148018837, -0.10913727432489395, 0.08549630641937256, -0.08915477991104126, 0.03328920155763626, 0.038260769098997116, -0.0951346829533577, -0.11748822033405304, -0.07748425751924515, 0.13025923073291779, 0.07970496267080307, -0.0018375783693045378, -0.010645721107721329, 0.07100877910852432, -0.005921723786741495, 0.07607460767030716, 0.25544750690460205, -0.31088972091674805, -0.03874797746539116, 0.19786159694194794, 0.03267781808972359, -0.055606402456760406, -0.0010495600290596485, 0.037377335131168365, -0.01061239093542099, -0.007985244505107403, 0.0429997518658638, -0.08162202686071396, 0.17307671904563904, -0.06016308441758156, -0.1192910373210907, 0.011027365922927856, 0.14200147986412048, -0.01833384484052658, -0.04222438484430313, -0.1035558357834816, -0.10804188996553421, 0.09956634789705276, -0.05887453258037567, -0.057062529027462006, 0.02349638193845749, -0.006517182104289532, 0.00647822255268693, -0.029438989236950874, -0.0399588905274868, -0.04118568077683449, -0.05687786638736725, 0.1231052577495575, 0.0023903786204755306, -0.002279575914144516, -0.005869485903531313, 0.08738602697849274, -0.1198456659913063, -0.11964918673038483, -0.013363020494580269, -0.07287956029176712, 0.0324099026620388, -0.010941755026578903, -0.05135035142302513, -0.11720024049282074, 0.07617011666297913, 0.2460498958826065, -0.061581820249557495, 0.06854994595050812, 0.07151197642087936, 0.04596604406833649, -0.0317060686647892, 0.0009703785763122141, -0.05785942077636719, -0.1865924447774887, 0.026493912562727928, 0.05492487549781799, 0.09048617631196976, -0.02317991852760315, -0.07088474184274673, -0.03918500244617462, 0.0532158724963665, -0.008379798382520676, 0.013928291387856007, 0.10173497349023819, -0.08189301937818527, -0.04155844449996948, 0.20403476059436798, -0.07965461909770966, -0.031215127557516098, -0.01823222078382969, -0.024241534993052483, 0.08296014368534088, 0.0858868882060051, 0.03991827368736267, 0.031033659353852272, 0.10381913185119629, -0.07330580800771713, -0.035618558526039124, -0.004170284606516361, -0.08332834392786026, 0.012277396395802498, -0.04151748865842819, -0.05012799799442291, -0.12400011718273163, -0.18452738225460052, -0.008143974468111992, 0.031488075852394104, -0.025952711701393127, 0.006152115762233734, -0.05357968807220459, -0.01774226687848568, 0.012745234183967113, 0.006103802938014269, -0.04489675536751747, -0.019798897206783295, 0.01830632984638214, 0.003939498215913773, 0.08931621164083481, -0.11600486189126968, 0.008969306014478207, -0.09173201024532318, 0.0997283086180687, -0.2662090063095093, 0.06496351957321167, -0.07147497683763504, 0.02673196978867054, -0.141739159822464, -0.020747551694512367, -0.015177300199866295, 0.027812357991933823, 0.02316272258758545, 0.15667970478534698, -0.11457616090774536, -0.10235193371772766, 0.17681346833705902, -0.13387419283390045, -0.11796706914901733, 0.050764795392751694, -0.008498354814946651, 0.08480825275182724, 0.10392023622989655, 0.19876503944396973, 0.05751394107937813, -0.047812603414058685, -0.043505847454071045, -0.016126418486237526, 0.07292468100786209, 0.032737210392951965, 0.09907123446464539, -0.042194899171590805, -0.06967279314994812, 0.032990846782922745, -0.09119284898042679, 0.029571503400802612, -0.0033947392366826534, -0.03825066238641739, -0.008843076415359974, -0.04557572305202484, 0.13930529356002808, -0.03746283799409866, 0.012941108085215092, -0.07828003168106079, -0.08859231323003769, 0.09789065271615982, 0.06489269435405731, -0.032025303691625595, -0.0005005280254408717, -0.0836712196469307, 0.09871434420347214, -0.034465473145246506, 0.025090552866458893, -0.10741916298866272, -0.11292627453804016, -0.021776126697659492, -0.03328566998243332, 0.01249708142131567, 0.005491104442626238, 0.08880837261676788, 0.056788042187690735, -0.07585711032152176, -0.02072548121213913, 0.10333473980426788, 0.022116590291261673, -0.01679714396595955, -0.17166630923748016, -0.06251664459705353, -0.052765458822250366, 0.25170111656188965, -0.08569198846817017, 0.06919628381729126, 0.03693750873208046, 0.18069671094417572, 0.012449851259589195, -0.015942338854074478, 0.038125164806842804, 0.011643790639936924, -0.01437212061136961, -0.011824097484350204, 0.11322887241840363, -0.03354957699775696, -0.1504383534193039, 0.09531936794519424, -0.16110366582870483, 0.12087453156709671, 0.10971896350383759, 0.017771996557712555, -0.0751522034406662, -0.06461193412542343, 0.015792574733495712, -0.05793413892388344, 0.08075287938117981, -0.0819726437330246, 0.0938219204545021, 0.016506658867001534, 0.09573931246995926, -0.07748076319694519, -0.016071267426013947, -0.004935798235237598, -0.04964040592312813, -0.012543952092528343, 0.07571897655725479, -0.075836680829525, -0.1609545648097992, 0.10192879289388657, 0.20159651339054108, -0.0024110437370836735, 0.11116137355566025, 0.03728751465678215, 0.02652883157134056, -0.05127433314919472, 0.034060146659612656, 0.013688583858311176, -0.004026644863188267, -0.09760045260190964, -0.008116398938000202, 0.04883900284767151, -0.006383879110217094, 0.017998022958636284, -0.08016156405210495, 0.010774757713079453, 0.02223236672580242, -0.009351869113743305, 0.051047053188085556, 0.08174993097782135, -0.016066960990428925, 0.07754972577095032, 0.048939697444438934, -0.07300654798746109, 0.023049067705869675, -0.004347160924226046, -0.06851489841938019, 0.18800120055675507, -0.07526993751525879, -0.19849644601345062, -0.13882561028003693, -0.11336401104927063, -0.05776934325695038, 0.01488581020385027, 0.040707387030124664, -0.00821052584797144, -0.023940013721585274, -0.09483464062213898, 0.13638804852962494, 0.051888782531023026, -0.023832429200410843, -0.010015247389674187, -0.037580881267786026, 0.01781025156378746, -0.10308004915714264, -0.03244581073522568, 0.00964854285120964, -0.06578974425792694, 0.05433155223727226, -0.09180012345314026, 0.06154206767678261, 0.10447661578655243, 0.028080351650714874, -0.021747060120105743, -0.042259715497493744, 0.2033211588859558, -0.030986959114670753, 0.1010715439915657, 0.14222611486911774, -0.028607405722141266, 0.06168926879763603, 0.15995153784751892, 0.03889816626906395, -0.03722088411450386, -0.008481021970510483, -0.024645976722240448, -0.06857045739889145, -0.1797235608100891, -0.09261789917945862, -0.011963541619479656, 0.09653990715742111, 0.053702063858509064, 0.016210516914725304, 0.10017957538366318, 0.10973479598760605, -0.0728863999247551, 0.011142893694341183, 0.10501153022050858, 0.08126599341630936, 0.1707884520292282, 0.011826438829302788, 0.15191572904586792, -0.01987701654434204, -0.043566200882196426, 0.034249357879161835, 0.021569736301898956, 0.01934773661196232, 0.017856750637292862, 0.13775786757469177, 0.0445164255797863, 0.053914133459329605, 0.0568980798125267, 0.07774618268013, -0.020766127854585648, -0.008754887618124485, -0.05370744690299034, -0.10556045174598694, -0.059869155287742615, 0.052329495549201965, -0.05094681680202484, 0.06850436329841614, -0.011604180559515953, -0.0013638666132465005, 0.03960398584604263, 0.11277013272047043, 0.09009063988924026, -0.2698194980621338, -0.10922403633594513, 0.056115832179784775, 0.05076875165104866, 0.00677691213786602, 0.015202243812382221, 0.029322968795895576, -0.09170451760292053, 0.16806288063526154, -0.05183333531022072, 0.05544427037239075, -0.044307321310043335, 0.029921879991889, -0.04832787066698074, 0.051170483231544495, -0.0055454629473388195, 0.033137887716293335, -0.15647709369659424, 0.09365775436162949, 0.03475392982363701, -0.021986493840813637, 0.021773304790258408, 0.019520113244652748, 0.0496613048017025, 0.09931046515703201, 0.07802098244428635, -0.0059050596319139, 0.0007088600541464984, -0.0730600357055664, -0.10248800367116928, -0.017842702567577362, 0.05457625910639763, -0.0963568314909935, 0.0888558104634285, -0.031108811497688293, -0.05785950645804405, 0.0034435095731168985, 0.10117750614881516, -0.1139930710196495, -0.09661976248025894, 0.0854998528957367, 0.0627027153968811, 0.06586838513612747, -0.09985961019992828, -0.06072091683745384, -0.053723882883787155, 0.17822203040122986, -0.13499073684215546, -0.10382668673992157, -0.07750917971134186, -0.07679026573896408, 0.11638932675123215, -0.08590735495090485, 0.0681837797164917, -0.04571918398141861, 0.04373173788189888, -0.056223854422569275, -0.11720260977745056, 0.10834332555532455, -0.07540078461170197, -0.1270349621772766, -0.03605863079428673, 0.18045301735401154, -0.026171168312430382, 0.040346406400203705, 0.007462585344910622, 0.06975069642066956, -0.018189631402492523, -0.04066852107644081, -0.007372009102255106, 0.06455178558826447, -0.03930382430553436, 0.08775229752063751, -0.01622275449335575, -0.11781463772058487, -0.02494136430323124, 0.0415426604449749, 0.1057964414358139, 0.26453015208244324, -0.026576576754450798, 0.09843339771032333, 0.14319583773612976, -0.02840392477810383, -0.22924384474754333, -0.07289975136518478, 0.002312966389581561, -0.015132123604416847, 0.030871080234646797, -0.11038349568843842, 0.0945277065038681, 0.13035999238491058, -0.0134115694090724, 0.0722341388463974, -0.3346579074859619, -0.12607131898403168, 0.0938020646572113, 0.07126495987176895, 0.19155047833919525, -0.14896321296691895, -0.10706204921007156, -0.08214366436004639, -0.23313894867897034, 0.07590997964143753, -0.09583009779453278, 0.06667115539312363, -0.032881367951631546, 0.038990117609500885, 0.006986340042203665, -0.03248047083616257, 0.14002107083797455, 0.021980008110404015, 0.012284188531339169, -0.07859539985656738, -0.0897207111120224, 0.11084475368261337, -0.049045663326978683, 0.062245212495326996, -0.13056549429893494, 0.02142598107457161, -0.010702583938837051, -0.03817405551671982, -0.06411884725093842, 0.07136346399784088, -0.05891634151339531, -0.05732440575957298, -0.024425743147730827, 0.02520711161196232, -0.004932665266096592, 0.030572200194001198, 0.21070393919944763, -0.05020565912127495, 0.1806785613298416, 0.18265388906002045, 0.08718278259038925, -0.08958941698074341, -0.04692647233605385, -0.01956642046570778, -0.05957872048020363, 0.04635949432849884, -0.1261998564004898, -0.016675865277647972, 0.0765485167503357, 0.025290722027420998, 0.08438677340745926, 0.03623293340206146, -0.04771570861339569, -0.021008096635341644, 0.07093913108110428, -0.1677773892879486, -0.14798696339130402, 0.006507662124931812, 0.06093493103981018, -0.04325129836797714, 0.11391519010066986, 0.23920121788978577, -0.03837841376662254, -0.020626911893486977, 0.03982073441147804, -0.001675912644714117, -0.050356440246105194, 0.15707656741142273, 0.006226871162652969, 0.06402856856584549, -0.09390304982662201, 0.02109445258975029, 0.033446915447711945, -0.041381753981113434, -0.038291845470666885, 0.05666545405983925, -0.14415989816188812, -0.07557740062475204, -0.14522694051265717, 0.12453772872686386, -0.08500189334154129, 0.001404422800987959, -0.09655405580997467, -0.13381651043891907, -0.0027105938643217087, 0.13399295508861542, 0.09569918364286423, 0.05357084423303604, 0.02243434451520443, -0.0627330094575882, 0.036118824034929276, 0.053670015186071396, 0.030121948570013046, 0.10077346861362457, -0.05980828031897545, 0.026350658386945724, -0.01953171379864216, -0.02165284752845764, -0.045289166271686554, 0.002184472978115082, -0.1509750336408615, -0.03290519118309021, -0.21536344289779663, 0.004950888454914093, -0.15125860273838043, -0.029590019956231117, -0.03927561640739441, 0.0133194699883461, 0.003747997572645545, -0.015087918378412724, -0.024611560627818108, -0.023533079773187637, -0.03662945702672005, 0.07180263102054596, -0.10625502467155457, 0.0016620609676465392, 0.03961095213890076, -0.0610833466053009, 0.05039878562092781, 0.03419620916247368, -0.009073406457901001, -0.00489542493596673, -0.12338148802518845, -0.014581697061657906, 0.03998950496315956, -0.03377566114068031, 0.01559498067945242, -0.10485735535621643, -0.04054964333772659, -0.01619148440659046, -0.014339392073452473, 0.027953103184700012, 0.11641217023134232, -0.07654356956481934, 0.051805440336465836, -0.028766226023435593, -0.04899217560887337, -0.0498364083468914, -0.006951083429157734, 0.1110144555568695, 0.023060200735926628, 0.14779841899871826, -0.0805656909942627, 0.03346749022603035, -0.1384742259979248, -0.013102582655847073, -0.0126724848523736, -0.14381656050682068, -0.06375475972890854, -0.043909668922424316, -0.0022548323031514883, -0.0212103258818388, 0.11238346248865128, -0.08576630055904388, -0.09716513752937317, 0.02529752627015114, 0.00477809552103281, 0.07852545380592346, 0.03213394805788994, 0.15056219696998596, 0.07737653702497482, -0.031010311096906662, -0.037172283977270126, 0.07603565603494644, 0.032107144594192505, 0.0053541758097708225, 0.07728078961372375, 0.14099779725074768, -0.000002636792487464845, 0.08290113508701324, 0.02469639852643013, 0.03797745704650879, -0.031552381813526154, 0.055464111268520355, 0.024929847568273544, 0.024281879886984825, -0.016584593802690506, 0.1721053272485733, 0.14232900738716125, -0.12619848549365997, 0.040970027446746826, -0.005140084307640791, -0.01002095639705658, -0.08092568069696426, -0.1134449690580368, -0.09128963202238083, -0.08639784902334213, -0.05692199617624283, -0.08226726204156876, -0.09542201459407806, 0.13592010736465454, -0.006630479823797941, 0.027982164174318314, 0.17722563445568085, -0.07574418187141418, -0.053087253123521805, 0.03463488072156906, 0.004365664441138506, -0.045925039798021317, 0.004153328016400337, -0.05391800403594971, 0.006001985166221857, 0.04297015815973282, -0.009149796329438686, 0.004598613828420639, 0.018801569938659668, -0.0041040536016225815, -0.01858222484588623, -0.13472607731819153, -0.011026574298739433, 0.06582905352115631, 0.004770041909068823, 0.00480677792802453, 0.007694635074585676, -0.038777802139520645, -0.018046148121356964, 0.11661631613969803, -0.01819024235010147, -0.08296913653612137, -0.032179128378629684, 0.15819501876831055, -0.02827404998242855, 0.066433846950531, 0.003303431672975421, -0.06437998265028, 0.024369921535253525, 0.09994911402463913, 0.3111204206943512, -0.03373263031244278, 0.023110559210181236, 0.015534786507487297, 0.019501959905028343, 0.04255768284201622, 0.029719674959778786, 0.005821031052619219, 0.15178243815898895, -0.036626577377319336, 0.0495661199092865, -0.0029624977614730597, -0.06800169497728348, -0.06371063739061356, -0.0167621411383152, 0.007029369007796049, -0.008075139485299587, -0.012787919491529465, 0.08515293896198273, -0.0845525711774826, -0.11287948489189148, 0.04535609856247902, -0.1895492970943451, -0.10954240709543228, -0.0820273607969284, 0.05860595032572746, 0.027088193222880363, 0.09965185075998306, -0.020607691258192062, -0.02887376770377159, 0.09302084892988205, -0.009882171638309956, -0.10221419483423233, -0.06261496245861053, 0.017576713114976883, -0.041226644068956375, 0.07746277749538422, -0.03897705674171448, 0.03620166704058647, 0.12073837965726852, -0.005332360975444317, -0.061021726578474045, 0.02605467289686203, 0.07730630785226822, -0.026722870767116547, 0.00522364629432559, 0.046450067311525345, -0.033297356218099594, 0.1857147216796875, 0.07422595471143723, -0.21186155080795288, 0.02816169522702694, 0.13871143758296967, -0.07170333713293076, -0.07313799858093262, 0.052640363574028015, -0.06394922733306885, 0.08854091167449951, 0.16382279992103577, -0.039402853697538376, -0.026215849444270134, -0.011317583732306957, 0.01940617524087429, 0.09714440256357193, 0.029511485248804092, -0.06711836904287338, -0.1818682849407196, -0.029285553842782974, -0.0432889461517334, 0.03490636870265007, -0.3052663207054138, -0.08469218015670776, -0.11598582565784454, 0.0074462671764194965, -0.07580012083053589, 0.06439150124788284, 0.07150955498218536, 0.03880058228969574, -0.004340976942330599, -0.08871690928936005, -0.015305601060390472, 0.12644962966442108, -0.09482832998037338, -0.10105627775192261 ]
null
null
null
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-base-finetuned-yoruba-lora-text-classification This model is a fine-tuned version of [Davlan/xlm-roberta-base-finetuned-yoruba](https://huggingface.co/Davlan/xlm-roberta-base-finetuned-yoruba) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.5843 - Precision: 0.6850 - Recall: 1.0 - F1 and accuracy: {'accuracy': 0.6915167095115681, 'f1': 0.8130841121495327} ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 and accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:----------------------------------------------------------:| | No log | 1.0 | 388 | 0.6314 | 0.6710 | 1.0 | {'accuracy': 0.6709511568123393, 'f1': 0.8030769230769231} | | 0.6474 | 2.0 | 776 | 0.6289 | 0.6710 | 1.0 | {'accuracy': 0.6709511568123393, 'f1': 0.8030769230769231} | | 0.652 | 3.0 | 1164 | 0.6270 | 0.6710 | 1.0 | {'accuracy': 0.6709511568123393, 'f1': 0.8030769230769231} | | 0.646 | 4.0 | 1552 | 0.6235 | 0.6710 | 1.0 | {'accuracy': 0.6709511568123393, 'f1': 0.8030769230769231} | | 0.646 | 5.0 | 1940 | 0.6188 | 0.6710 | 1.0 | {'accuracy': 0.6709511568123393, 'f1': 0.8030769230769231} | | 0.6459 | 6.0 | 2328 | 0.6092 | 0.6710 | 1.0 | {'accuracy': 0.6709511568123393, 'f1': 0.8030769230769231} | | 0.6281 | 7.0 | 2716 | 0.5958 | 0.6710 | 1.0 | {'accuracy': 0.6709511568123393, 'f1': 0.8030769230769231} | | 0.6268 | 8.0 | 3104 | 0.5912 | 0.6727 | 1.0 | {'accuracy': 0.6735218508997429, 'f1': 0.8043143297380586} | | 0.6268 | 9.0 | 3492 | 0.5861 | 0.6850 | 1.0 | {'accuracy': 0.6915167095115681, 'f1': 0.8130841121495327} | | 0.6105 | 10.0 | 3880 | 0.5843 | 0.6850 | 1.0 | {'accuracy': 0.6915167095115681, 'f1': 0.8130841121495327} | ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.1
{"tags": ["generated_from_trainer"], "metrics": ["precision", "recall"], "base_model": "Davlan/xlm-roberta-base-finetuned-yoruba", "model-index": [{"name": "xlm-roberta-base-finetuned-yoruba-lora-text-classification", "results": []}]}
null
nhankins/yo_xlmroberta_lora_adapter_1.0
[ "tensorboard", "safetensors", "generated_from_trainer", "base_model:Davlan/xlm-roberta-base-finetuned-yoruba", "region:us" ]
2024-02-10T15:55:27+00:00
[]
[]
TAGS #tensorboard #safetensors #generated_from_trainer #base_model-Davlan/xlm-roberta-base-finetuned-yoruba #region-us
xlm-roberta-base-finetuned-yoruba-lora-text-classification ========================================================== This model is a fine-tuned version of Davlan/xlm-roberta-base-finetuned-yoruba on the None dataset. It achieves the following results on the evaluation set: * Loss: 0.5843 * Precision: 0.6850 * Recall: 1.0 * F1 and accuracy: {'accuracy': 0.6915167095115681, 'f1': 0.8130841121495327} Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 1e-05 * train\_batch\_size: 4 * eval\_batch\_size: 4 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 10 ### Training results ### Framework versions * Transformers 4.35.2 * Pytorch 2.1.0+cu121 * Datasets 2.17.0 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10", "### Training results", "### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ "TAGS\n#tensorboard #safetensors #generated_from_trainer #base_model-Davlan/xlm-roberta-base-finetuned-yoruba #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10", "### Training results", "### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ 45, 98, 4, 33 ]
[ "passage: TAGS\n#tensorboard #safetensors #generated_from_trainer #base_model-Davlan/xlm-roberta-base-finetuned-yoruba #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ -0.10487453639507294, -0.0021084982436150312, -0.0014557790709659457, 0.0918414443731308, 0.2006765753030777, 0.027966652065515518, 0.14998400211334229, 0.062311794608831406, -0.09077467024326324, 0.057505566626787186, 0.10070547461509705, 0.10529245436191559, 0.013317002914845943, 0.08935552835464478, -0.059268053621053696, -0.19616760313510895, -0.009650140069425106, 0.025079699233174324, -0.08813221007585526, 0.11338666081428528, 0.07100605964660645, -0.16016431152820587, 0.07486005872488022, -0.040587667375802994, -0.2568352222442627, 0.05070657655596733, 0.03259272873401642, -0.055375225841999054, 0.11728452146053314, -0.007729213219136, 0.17856882512569427, 0.015256247483193874, 0.09902814030647278, -0.17177420854568481, 0.023678604513406754, 0.07395381480455399, 0.005086051300168037, 0.05925159156322479, 0.052391938865184784, -0.018181608989834785, 0.06312216818332672, -0.13610795140266418, 0.037744730710983276, 0.010421493090689182, -0.15275166928768158, -0.2104194313287735, -0.08338162302970886, -0.02045634388923645, 0.078050397336483, 0.08081317692995071, -0.02377309836447239, 0.17756301164627075, -0.05421997606754303, 0.07995137572288513, 0.2545807659626007, -0.26242655515670776, -0.08904273808002472, 0.09446613490581512, 0.020425040274858475, 0.11689101904630661, -0.11267675459384918, -0.00009634196612751111, 0.09668296575546265, 0.03757258132100105, 0.09282571822404861, -0.039396919310092926, -0.07392316311597824, 0.016460474580526352, -0.15297704935073853, 0.043763112276792526, 0.08197717368602753, 0.03734614700078964, -0.04541540518403053, 0.02311933971941471, -0.07250921428203583, -0.11966452747583389, -0.06557495146989822, -0.04044220596551895, 0.05003376305103302, -0.06338662654161453, -0.09860700368881226, -0.012295416556298733, -0.09010601043701172, -0.07933693379163742, -0.05459725856781006, 0.2656770348548889, 0.049679841846227646, 0.017819270491600037, -0.051963016390800476, 0.10301419347524643, -0.05809246748685837, -0.14385150372982025, 0.04735409468412399, 0.03615737706422806, -0.01203232817351818, -0.0619930699467659, -0.059378426522016525, -0.1294003278017044, 0.04451725631952286, 0.10385902971029282, -0.09535735845565796, 0.04080304130911827, 0.026571182534098625, 0.03229031339287758, -0.11092216521501541, 0.1483188271522522, -0.05929048731923103, -0.03280172869563103, 0.04328235611319542, 0.08132342249155045, 0.032985907047986984, 0.0020035302732139826, -0.07852772623300552, 0.014668828807771206, 0.09384070336818695, 0.02036607638001442, -0.09032703936100006, 0.03626169264316559, -0.03693016991019249, 0.010955512523651123, -0.04054225608706474, -0.11626823991537094, 0.0289918202906847, -0.004990281071513891, -0.05827615410089493, -0.03971053659915924, 0.0025739616248756647, 0.021883612498641014, 0.03468097373843193, 0.09823162108659744, -0.09072894603013992, 0.04537142813205719, -0.10885294526815414, -0.11073583364486694, -0.017343437299132347, -0.05333317071199417, 0.03445423021912575, -0.10332837700843811, -0.15869945287704468, -0.02314371056854725, 0.039319757372140884, -0.02961842343211174, 0.012659700587391853, -0.044968001544475555, -0.09806878119707108, -0.01831117831170559, -0.000370582944015041, 0.1269131451845169, -0.05390411242842674, 0.11311838030815125, 0.09797286242246628, 0.06826843321323395, -0.06075042858719826, 0.010598903521895409, -0.08777142316102982, 0.027219656854867935, -0.2318926751613617, 0.00871064979583025, -0.05506845563650131, 0.06553327292203903, -0.0576418898999691, -0.06525517255067825, -0.031625088304281235, 0.009248467162251472, 0.11079061031341553, 0.1080591157078743, -0.18925315141677856, -0.0688643679022789, 0.17820578813552856, -0.09882999956607819, -0.10580895841121674, 0.10686230659484863, -0.055091239511966705, 0.051096487790346146, 0.077351875603199, 0.18746452033519745, -0.017075814306735992, -0.12862609326839447, -0.003643396543338895, -0.0646757259964943, 0.06044293940067291, -0.035485316067934036, 0.03710542246699333, 0.009176628664135933, -0.016476303339004517, 0.016771769151091576, -0.03423822298645973, 0.05336494743824005, -0.12093639373779297, -0.07324501872062683, -0.052342966198921204, -0.11425811797380447, 0.036140188574790955, 0.07325530052185059, 0.059582192450761795, -0.12501588463783264, -0.048965249210596085, 0.08828941732645035, 0.0788540244102478, -0.06624232977628708, 0.011929119937121868, -0.057110633701086044, 0.07318582385778427, -0.115212082862854, -0.048453640192747116, -0.1532687544822693, -0.07753763347864151, -0.0026986703742295504, 0.04194629564881325, 0.03242094814777374, -0.03992980718612671, 0.09522955864667892, 0.08244595676660538, -0.07316964864730835, -0.01975683495402336, -0.05766618624329567, -0.0009054510737769306, -0.12799616158008575, -0.19305606186389923, 0.008497175760567188, -0.023824725300073624, 0.09229005873203278, -0.23150727152824402, 0.0386778898537159, -0.052956126630306244, 0.10607830435037613, 0.04537772759795189, -0.01798098348081112, -0.024787645787000656, 0.10317149758338928, 0.0028403536416590214, -0.07052589952945709, 0.01980501227080822, -0.03153681382536888, -0.07954476773738861, -0.08680974692106247, -0.13544772565364838, 0.19046340882778168, 0.1367599219083786, -0.09786448627710342, -0.08635327219963074, 0.042694203555583954, -0.040947455912828445, -0.015341914258897305, -0.05835985764861107, 0.05099864676594734, 0.12295558303594589, -0.010033056139945984, 0.1200757622718811, -0.07936406135559082, -0.015385186299681664, 0.016249818727374077, -0.05939076840877533, 0.05895324796438217, 0.09153066575527191, 0.130243182182312, -0.05560412257909775, 0.12049486488103867, 0.16110260784626007, -0.12645570933818817, 0.13783909380435944, -0.056596703827381134, -0.07961682975292206, -0.02786489576101303, -0.002800285117700696, 0.004537518136203289, 0.17375053465366364, -0.04621671885251999, 0.02405952289700508, -0.016170701012015343, -0.002753298729658127, 0.029497411102056503, -0.22879812121391296, -0.06624995172023773, 0.004156551323831081, -0.026804842054843903, 0.013516515493392944, -0.0031693866476416588, 0.013232861645519733, 0.1073303297162056, -0.047812048345804214, -0.05366505682468414, 0.007026879116892815, 0.002649680245667696, -0.06269049644470215, 0.21855488419532776, -0.06081556901335716, -0.031224682927131653, -0.06028416007757187, -0.03984282538294792, -0.05122019350528717, 0.0007852944545447826, 0.03380493447184563, -0.13449101150035858, -0.026543425396084785, -0.07039794325828552, 0.026360202580690384, 0.04639625921845436, 0.049611736088991165, 0.032178062945604324, -0.006260013673454523, 0.08084144443273544, -0.11232215911149979, 0.01561168022453785, -0.0801970437169075, -0.08301331847906113, 0.059955183416604996, 0.07968324422836304, 0.1287509799003601, 0.15098103880882263, -0.02685665525496006, -0.003944889642298222, -0.01566489413380623, 0.2620628774166107, -0.08434467017650604, -0.03961214795708656, 0.09649593383073807, -0.01007149275392294, 0.03759971633553505, 0.12092569470405579, 0.0825352594256401, -0.13239888846874237, 0.017482860013842583, 0.05553659051656723, -0.037791963666677475, -0.21143823862075806, -0.013439078815281391, -0.023047635331749916, -0.08213634043931961, 0.031660500913858414, 0.02836652658879757, -0.019346831366419792, 0.07671260088682175, 0.0422687828540802, 0.0769084170460701, -0.05258578062057495, 0.0452050119638443, 0.008135265670716763, 0.05632094666361809, 0.09412028640508652, -0.0534677654504776, -0.08493044972419739, 0.013411853462457657, -0.016211530193686485, 0.22482293844223022, 0.009862390346825123, 0.051914576441049576, 0.07503966987133026, 0.16817133128643036, -0.004016564693301916, 0.06607308983802795, 0.003253993345424533, -0.08277646452188492, 0.012181147001683712, -0.05571179836988449, 0.022307243198156357, -0.0029742498882114887, -0.11910942941904068, 0.09156814217567444, -0.08549348264932632, 0.0085876090452075, 0.07499717175960541, 0.151697039604187, 0.01804514229297638, -0.3078320026397705, -0.050004225224256516, 0.012750512920320034, 0.002367856679484248, -0.01136326789855957, 0.00043742882553488016, 0.14682404696941376, -0.018133973702788353, 0.03480016812682152, -0.06882064789533615, 0.05835331231355667, 0.04224693402647972, 0.02913622558116913, 0.015715794637799263, 0.1707688421010971, -0.035397954285144806, 0.016382139176130295, -0.28968116641044617, 0.274901419878006, 0.036839261651039124, 0.13022395968437195, -0.014388863928616047, -0.02786010131239891, 0.02909877337515354, 0.05417770519852638, 0.02926761470735073, -0.02175571583211422, -0.039194174110889435, -0.2027461975812912, -0.03533908724784851, 0.07047781348228455, 0.15464653074741364, 0.007044235244393349, 0.12310203909873962, 0.019709015265107155, 0.015923356637358665, 0.08301332592964172, -0.05896403640508652, -0.15799039602279663, -0.036799024790525436, -0.05052116513252258, 0.022707674652338028, -0.07538556307554245, -0.0784941166639328, -0.0959600955247879, -0.14971478283405304, 0.08317694813013077, 0.01165817491710186, -0.015807965770363808, -0.10625293105840683, 0.09939387440681458, 0.09867789596319199, -0.06253320723772049, 0.029723407700657845, 0.03958071023225784, 0.038939472287893295, 0.02362048253417015, -0.04225169122219086, 0.09600038081407547, -0.06007400527596474, -0.15905529260635376, -0.06480581313371658, 0.08095259964466095, 0.048484016209840775, 0.03712534159421921, -0.008325453847646713, 0.009561981074512005, 0.004392150789499283, -0.08723714202642441, 0.04848853871226311, -0.04263608157634735, 0.03495867922902107, 0.04480331391096115, -0.03812949359416962, -0.017553487792611122, -0.06613202393054962, -0.031207263469696045, 0.11427971720695496, 0.30777058005332947, -0.050811510533094406, -0.05033498257398605, 0.042652565985918045, -0.06195669248700142, -0.16771742701530457, 0.11035079509019852, 0.05159345641732216, 0.02399400807917118, 0.0762263610959053, -0.12354688346385956, 0.15359610319137573, 0.11495552211999893, -0.004411834292113781, 0.1426638662815094, -0.3127264976501465, -0.1404912918806076, 0.08912043273448944, 0.20572496950626373, 0.16997100412845612, -0.16999505460262299, -0.01790524087846279, -0.021930696442723274, -0.08614873886108398, 0.07829514890909195, -0.196383535861969, 0.09877625852823257, -0.010009784251451492, 0.07689616829156876, 0.005371957551687956, -0.0567438006401062, 0.14596867561340332, 0.00256381300278008, 0.15152955055236816, -0.04195237532258034, -0.03107352741062641, 0.07818520069122314, -0.028661830350756645, -0.03033500723540783, -0.03553329035639763, 0.018814027309417725, -0.012205294333398342, -0.01928171142935753, -0.06754978746175766, 0.054596755653619766, -0.03639639914035797, -0.06034856662154198, -0.041259706020355225, 0.016719838604331017, 0.0036420554388314486, -0.023477425798773766, 0.10810410976409912, 0.022736690938472748, 0.17915847897529602, 0.059662897139787674, 0.04835675656795502, -0.062223322689533234, 0.0030629108659923077, 0.03680841252207756, -0.0195884108543396, 0.06049090251326561, -0.14581604301929474, 0.02052294835448265, 0.10655348747968674, 0.02174239605665207, 0.10367125272750854, 0.06903215497732162, -0.0627249926328659, 0.04561273753643036, 0.08987350016832352, -0.1512545943260193, -0.1007615178823471, 0.015510311350226402, -0.0967026874423027, -0.08609456568956375, 0.09180103242397308, 0.09533637017011642, -0.06443437933921814, -0.01677005924284458, -0.028732215985655785, -0.011724126525223255, -0.07558900862932205, 0.22267085313796997, 0.11124633252620697, 0.03386668488383293, -0.08268516510725021, 0.08746964484453201, 0.03123965859413147, -0.04647593945264816, -0.004166920203715563, 0.05996817722916603, -0.0646653100848198, -0.010470641776919365, 0.11109110713005066, 0.2522668242454529, -0.031936295330524445, -0.03826064243912697, -0.1741223931312561, -0.08608843386173248, -0.002308879280462861, 0.21676836907863617, 0.09053212404251099, -0.03371444717049599, 0.01508641242980957, 0.03541858121752739, -0.16109973192214966, 0.08495156466960907, 0.06130390986800194, 0.09943728148937225, -0.14687609672546387, 0.16329419612884521, 0.009895531460642815, 0.01712978444993496, -0.033365242183208466, 0.05929961055517197, -0.12121644616127014, 0.024408141151070595, -0.1269616186618805, -0.04137421026825905, 0.00274465698748827, -0.02006184309720993, 0.013460181653499603, -0.08402876555919647, -0.09906236082315445, 0.019657747820019722, -0.12274827063083649, 0.00010214754001935944, 0.061971988528966904, 0.017732057720422745, -0.10646301507949829, -0.03637375310063362, 0.025281868875026703, -0.048842981457710266, 0.02572568878531456, 0.04081162065267563, 0.03225861117243767, 0.09420540928840637, -0.22736641764640808, 0.010210435837507248, 0.07661282271146774, -0.017317747697234154, 0.06132800132036209, -0.02844478003680706, -0.014333750121295452, -0.02031007409095764, 0.11151529848575592, 0.015798155218362808, 0.08365780860185623, -0.11862663924694061, 0.010813899338245392, -0.0704292580485344, -0.050776753574609756, -0.04118259996175766, 0.002521041315048933, 0.06556626409292221, 0.01124534197151661, 0.16848184168338776, -0.10267886519432068, 0.01998162269592285, -0.22099965810775757, -0.015250247903168201, -0.02539769932627678, -0.08311846852302551, -0.10452041029930115, -0.021164894104003906, 0.07848172634840012, -0.05994642525911331, 0.137558713555336, 0.002927491907030344, 0.03629061579704285, 0.011220287531614304, -0.056058675050735474, 0.014767156913876534, 0.03086206130683422, 0.24911683797836304, 0.04296766221523285, -0.014573058113455772, 0.03579692915081978, 0.06260653585195541, 0.11745492368936539, 0.05808640643954277, 0.188129261136055, 0.1829889565706253, -0.05067971348762512, 0.10867206752300262, 0.021810030564665794, -0.06812047213315964, -0.06291957199573517, 0.05519159510731697, -0.07349314540624619, 0.025311458855867386, 0.0005647318321280181, 0.18981072306632996, 0.13473933935165405, -0.1457291841506958, 0.014459694735705853, -0.03618156909942627, -0.08536237478256226, -0.10418420284986496, 0.02697526291012764, -0.09428359568119049, -0.16171520948410034, 0.03937389329075813, -0.1151992678642273, -0.004025489091873169, 0.13960136473178864, 0.0005210849922150373, 0.007683238014578819, 0.2424290031194687, 0.0894087627530098, 0.03909526392817497, 0.041612498462200165, 0.005603300407528877, -0.04614061489701271, -0.049475666135549545, -0.08036056160926819, -0.007631656248122454, -0.045550551265478134, 0.02143065258860588, -0.06310112029314041, -0.07487507164478302, 0.05491005629301071, -0.023548835888504982, -0.10050315409898758, 0.018786512315273285, 0.03724222630262375, 0.059364330023527145, 0.009046507067978382, 0.013746502809226513, 0.003474864410236478, -0.020541056990623474, 0.20946672558784485, -0.05105654150247574, -0.09047169983386993, -0.07009337842464447, 0.20909376442432404, 0.03931815177202225, 0.0078110843896865845, -0.01166567299515009, -0.11378128826618195, 0.05151721462607384, 0.19178342819213867, 0.16093969345092773, -0.11294014006853104, 0.006881874520331621, -0.04052451252937317, -0.015203083865344524, -0.09676260501146317, 0.10065270215272903, 0.10393738746643066, 0.005697646643966436, -0.10379571467638016, -0.05437755584716797, -0.06387443840503693, -0.005360707174986601, -0.056436412036418915, 0.029455747455358505, 0.04590260237455368, 0.018835868686437607, -0.06069813296198845, 0.07937488704919815, -0.041968658566474915, -0.14108921587467194, 0.09459108859300613, -0.1875966489315033, -0.15018583834171295, -0.03403306379914284, 0.09389562904834747, -0.011756799183785915, 0.05173962935805321, -0.05643871799111366, -0.018974851816892624, 0.0225821565836668, -0.027365582063794136, -0.04537773132324219, -0.09325805306434631, 0.06332946568727493, -0.12864412367343903, 0.22021746635437012, -0.027445495128631592, 0.08147569745779037, 0.1142379641532898, 0.029960406944155693, -0.0561564564704895, 0.10748427361249924, 0.04055401310324669, -0.0933690220117569, 0.005243554245680571, 0.12200131267309189, -0.047353345900774, 0.05637199804186821, 0.0454389825463295, -0.09228479862213135, 0.020973971113562584, -0.03554690256714821, -0.07580433040857315, -0.05995602533221245, -0.05920660123229027, -0.07709493488073349, 0.10959125310182571, 0.1656164675951004, -0.021538129076361656, 0.06160331517457962, -0.07005499303340912, 0.05606505274772644, 0.08828675746917725, 0.08337227255105972, -0.05225047841668129, -0.25794273614883423, 0.04091339558362961, 0.13632027804851532, -0.055248700082302094, -0.24156275391578674, -0.05588562786579132, 0.005689277779310942, -0.0647570788860321, -0.07625395059585571, 0.07939359545707703, 0.1361124962568283, 0.06378325074911118, -0.04600473493337631, -0.17382240295410156, -0.0822148248553276, 0.17054034769535065, -0.1382390260696411, -0.10255807638168335 ]
null
null
null
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # xlm-roberta-large-lora-text-classification This model is a fine-tuned version of [FacebookAI/xlm-roberta-large](https://huggingface.co/FacebookAI/xlm-roberta-large) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.6343 - Precision: 0.6371 - Recall: 0.9958 - F1 and accuracy: {'accuracy': 0.6353887399463807, 'f1': 0.7770491803278688} ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 1e-05 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 and accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:----------------------------------------------------------:| | No log | 1.0 | 372 | 0.6553 | 0.6381 | 1.0 | {'accuracy': 0.6380697050938338, 'f1': 0.779050736497545} | | 0.6758 | 2.0 | 744 | 0.6534 | 0.6381 | 1.0 | {'accuracy': 0.6380697050938338, 'f1': 0.779050736497545} | | 0.6765 | 3.0 | 1116 | 0.6518 | 0.6381 | 1.0 | {'accuracy': 0.6380697050938338, 'f1': 0.779050736497545} | | 0.6765 | 4.0 | 1488 | 0.6519 | 0.6381 | 1.0 | {'accuracy': 0.6380697050938338, 'f1': 0.779050736497545} | | 0.6722 | 5.0 | 1860 | 0.6470 | 0.6381 | 1.0 | {'accuracy': 0.6380697050938338, 'f1': 0.779050736497545} | | 0.6654 | 6.0 | 2232 | 0.6425 | 0.6381 | 1.0 | {'accuracy': 0.6380697050938338, 'f1': 0.779050736497545} | | 0.6626 | 7.0 | 2604 | 0.6419 | 0.6381 | 1.0 | {'accuracy': 0.6380697050938338, 'f1': 0.779050736497545} | | 0.6626 | 8.0 | 2976 | 0.6404 | 0.6371 | 0.9958 | {'accuracy': 0.6353887399463807, 'f1': 0.7770491803278688} | | 0.6547 | 9.0 | 3348 | 0.6356 | 0.6381 | 1.0 | {'accuracy': 0.6380697050938338, 'f1': 0.779050736497545} | | 0.6544 | 10.0 | 3720 | 0.6343 | 0.6371 | 0.9958 | {'accuracy': 0.6353887399463807, 'f1': 0.7770491803278688} | ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.1
{"license": "mit", "tags": ["generated_from_trainer"], "metrics": ["precision", "recall"], "base_model": "FacebookAI/xlm-roberta-large", "model-index": [{"name": "xlm-roberta-large-lora-text-classification", "results": []}]}
null
nhankins/es_xlmroberta_lora_adapter_1.0
[ "tensorboard", "safetensors", "generated_from_trainer", "base_model:FacebookAI/xlm-roberta-large", "license:mit", "region:us" ]
2024-02-10T15:58:41+00:00
[]
[]
TAGS #tensorboard #safetensors #generated_from_trainer #base_model-FacebookAI/xlm-roberta-large #license-mit #region-us
xlm-roberta-large-lora-text-classification ========================================== This model is a fine-tuned version of FacebookAI/xlm-roberta-large on the None dataset. It achieves the following results on the evaluation set: * Loss: 0.6343 * Precision: 0.6371 * Recall: 0.9958 * F1 and accuracy: {'accuracy': 0.6353887399463807, 'f1': 0.7770491803278688} Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 1e-05 * train\_batch\_size: 4 * eval\_batch\_size: 4 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 10 ### Training results ### Framework versions * Transformers 4.35.2 * Pytorch 2.1.0+cu121 * Datasets 2.17.0 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10", "### Training results", "### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ "TAGS\n#tensorboard #safetensors #generated_from_trainer #base_model-FacebookAI/xlm-roberta-large #license-mit #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10", "### Training results", "### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ 43, 98, 4, 33 ]
[ "passage: TAGS\n#tensorboard #safetensors #generated_from_trainer #base_model-FacebookAI/xlm-roberta-large #license-mit #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 1e-05\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ -0.11411944776773453, -0.014354150742292404, -0.0007589488523080945, 0.09452363848686218, 0.1862066090106964, 0.014524498023092747, 0.14500093460083008, 0.06619085371494293, -0.07135139405727386, 0.04561713710427284, 0.10388768464326859, 0.11677480489015579, 0.004332253709435463, 0.12270796298980713, -0.06765582412481308, -0.19440124928951263, -0.011587582528591156, 0.01763593591749668, -0.07966632395982742, 0.10960930585861206, 0.0799538791179657, -0.16385039687156677, 0.0838908851146698, -0.014964235946536064, -0.26626449823379517, 0.04923732578754425, 0.052088018506765366, -0.051715120673179626, 0.13924400508403778, 0.003040389623492956, 0.16612711548805237, 0.007665787823498249, 0.11875001341104507, -0.1484595090150833, 0.020173193886876106, 0.08163057267665863, -0.007792291231453419, 0.06912142038345337, 0.06650301814079285, -0.0021347813308238983, 0.09428755193948746, -0.12319101393222809, 0.06368310004472733, 0.020466171205043793, -0.13812638819217682, -0.24982939660549164, -0.08483570069074631, -0.000999231357127428, 0.07412204891443253, 0.09139060229063034, -0.021224230527877808, 0.1822977513074875, -0.08221816271543503, 0.08747369796037674, 0.2561880648136139, -0.2761102616786957, -0.08753178268671036, 0.06681518256664276, 0.020097577944397926, 0.09984526038169861, -0.11824743449687958, 0.002407018095254898, 0.10018373280763626, 0.044638391584157944, 0.11865317076444626, -0.031706228852272034, -0.09935476630926132, 0.018535681068897247, -0.14719943702220917, 0.03277745842933655, 0.07483352720737457, 0.04592499881982803, -0.039442453533411026, 0.004313264507800341, -0.05828340724110603, -0.0990181416273117, -0.05623513460159302, -0.048958487808704376, 0.06478915363550186, -0.06340636312961578, -0.10721219331026077, -0.0039034744258970022, -0.11247853189706802, -0.08339398354291916, -0.04753030091524124, 0.19537663459777832, 0.03880034387111664, 0.03292159363627434, -0.0512220524251461, 0.0929282158613205, -0.05080687254667282, -0.13048334419727325, 0.05251704901456833, 0.033719319850206375, -0.01889411173760891, -0.06580537557601929, -0.06348006427288055, -0.09456729143857956, 0.03047069162130356, 0.08931411057710648, -0.12200409919023514, 0.041307877749204636, 0.027705594897270203, 0.04344629868865013, -0.10250145941972733, 0.14911143481731415, -0.049329716712236404, -0.001316439243964851, 0.018185127526521683, 0.07387028634548187, 0.011915232054889202, 0.003833397990092635, -0.08357695490121841, 0.038734011352062225, 0.08633925765752792, 0.003028780920431018, -0.09125387668609619, 0.03539057821035385, -0.03417907655239105, 0.018781060352921486, -0.011824275366961956, -0.09454374015331268, 0.05312972143292427, 0.01165839284658432, -0.07234178483486176, -0.0392632894217968, 0.0039829835295677185, 0.021994780749082565, 0.03513052314519882, 0.12697501480579376, -0.09306316077709198, 0.05656483396887779, -0.11242559552192688, -0.11163872480392456, -0.007841184735298157, -0.0241798497736454, 0.02085241489112377, -0.10337428003549576, -0.14510034024715424, -0.02286067046225071, 0.048399120569229126, -0.028361953794956207, 0.013567382469773293, -0.03189215809106827, -0.09474509209394455, -0.014245495200157166, -0.016634760424494743, 0.14583668112754822, -0.05777449160814285, 0.09573648124933243, 0.06484492868185043, 0.06246025115251541, -0.09164852648973465, 0.02099337801337242, -0.09913821518421173, 0.003996932879090309, -0.25304245948791504, -0.0038938475772738457, -0.06304778158664703, 0.07100450992584229, -0.04680301994085312, -0.08164384961128235, -0.0042357174679636955, 0.016093138605356216, 0.1016293540596962, 0.08457235991954803, -0.21530216932296753, -0.08085174858570099, 0.14148692786693573, -0.100277379155159, -0.12116607278585434, 0.10689552128314972, -0.05726809427142143, 0.05892627686262131, 0.0903000608086586, 0.1876947581768036, -0.01049928180873394, -0.13884292542934418, -0.008286373689770699, -0.04951762035489082, 0.031603094190359116, -0.06592779606580734, 0.02584861032664776, 0.006695163436233997, -0.0055436622351408005, 0.017364535480737686, -0.03861396759748459, 0.039296235889196396, -0.12593568861484528, -0.07491492480039597, -0.056786224246025085, -0.10976359248161316, 0.013870671391487122, 0.0743442252278328, 0.0664726123213768, -0.13327395915985107, -0.05316741764545441, 0.10344962775707245, 0.06036173924803734, -0.0462527796626091, 0.009633494541049004, -0.04780009016394615, 0.06811395287513733, -0.05967449024319649, -0.04407863691449165, -0.16525432467460632, -0.07107904553413391, 0.0013265236048027873, 0.007212555501610041, 0.030637042596936226, 0.01021777093410492, 0.08789224922657013, 0.08491578698158264, -0.06881322711706161, -0.009435699321329594, -0.05890844762325287, 0.004740949720144272, -0.13243423402309418, -0.22561706602573395, -0.0005818906938657165, -0.022986512631177902, 0.07745640724897385, -0.24240532517433167, 0.03386382386088371, -0.03277420625090599, 0.0922727882862091, 0.02465013600885868, -0.00920097529888153, -0.07039714604616165, 0.08748675137758255, 0.009366972371935844, -0.05521994084119797, 0.038127318024635315, -0.019272862002253532, -0.06373214721679688, -0.07710114866495132, -0.10914155095815659, 0.20780862867832184, 0.14998935163021088, -0.11465970426797867, -0.09570711851119995, 0.022418880835175514, -0.06392935663461685, -0.011179251596331596, -0.0773061141371727, 0.04022643342614174, 0.16117678582668304, -0.012113964185118675, 0.12118956446647644, -0.08467891812324524, -0.03144673630595207, 0.014277276583015919, -0.04894457012414932, 0.05953848734498024, 0.08657639473676682, 0.14472724497318268, -0.08114106208086014, 0.11564001441001892, 0.16883081197738647, -0.10258499532938004, 0.12265323847532272, -0.05167720466852188, -0.06707067787647247, -0.021007677540183067, -0.00192382768727839, -0.009292062371969223, 0.18521109223365784, -0.04683380573987961, 0.01654638908803463, -0.021465912461280823, 0.0046907407231628895, 0.030629120767116547, -0.24335965514183044, -0.06224928796291351, -0.005755582824349403, -0.032226692885160446, -0.012601995840668678, -0.026399821043014526, 0.02894926257431507, 0.11985984444618225, -0.03971770405769348, -0.07497593760490417, 0.007187314331531525, 0.0038028424605727196, -0.07798925787210464, 0.2190164178609848, -0.05581061542034149, -0.06218797713518143, -0.06234806776046753, -0.007214652840048075, -0.03879634663462639, -0.009759019128978252, 0.046633921563625336, -0.1350557655096054, -0.029621731489896774, -0.07044698297977448, 0.023545654490590096, 0.0743241235613823, 0.045601580291986465, 0.02296166494488716, 0.005332197528332472, 0.09584128111600876, -0.11885173618793488, -0.0012855519307777286, -0.09399522095918655, -0.08074867725372314, 0.036421038210392, 0.09611041843891144, 0.11918770521879196, 0.13560697436332703, -0.03860729560256004, -0.0029225945472717285, -0.0255853570997715, 0.25425204634666443, -0.07743596285581589, -0.038694705814123154, 0.10103438794612885, 0.0017223161412402987, 0.04231837019324303, 0.11440730839967728, 0.09796302765607834, -0.13088883459568024, 0.008771643042564392, 0.05377458408474922, -0.028368311002850533, -0.2098608911037445, -0.02759740687906742, -0.01400752179324627, -0.06454309076070786, 0.05024396628141403, 0.03131189942359924, 0.0019840921740978956, 0.055585261434316635, 0.050005458295345306, 0.06165566295385361, -0.05171800032258034, 0.03934497386217117, 0.01390159409493208, 0.04947199672460556, 0.11468803882598877, -0.05198642238974571, -0.0734696239233017, 0.02934030070900917, -0.0535617358982563, 0.24495328962802887, 0.006011846475303173, 0.014139074832201004, 0.08522511273622513, 0.16653473675251007, -0.017097903415560722, 0.07218728959560394, -0.0017288256203755736, -0.09536072611808777, 0.006828194949775934, -0.05614153668284416, 0.031782668083906174, 0.021230477839708328, -0.11006606370210648, 0.0744599923491478, -0.11572875827550888, 0.0008891330799087882, 0.08087990432977676, 0.1995643824338913, 0.026187019422650337, -0.3159870505332947, -0.04704687371850014, -0.0058114854618906975, -0.0022559117060154676, -0.0008763917721807957, 0.008203876204788685, 0.15984952449798584, -0.02887471206486225, 0.039789967238903046, -0.064040906727314, 0.06932660937309265, 0.04232402145862579, 0.049362123012542725, 0.058138687163591385, 0.15268076956272125, -0.026090720668435097, 0.03121139481663704, -0.27708694338798523, 0.2983222007751465, 0.02931121736764908, 0.13219282031059265, -0.03758038580417633, -0.0251654963940382, 0.02632274106144905, 0.051783688366413116, 0.03407520800828934, -0.026724815368652344, -0.07234979420900345, -0.20753730833530426, -0.02925029769539833, 0.06979073584079742, 0.13051366806030273, 0.017014170065522194, 0.09562111645936966, 0.006472620647400618, 0.00971080269664526, 0.09597445279359818, -0.04822717234492302, -0.15547314286231995, -0.035666074603796005, -0.07339987903833389, 0.0295257568359375, -0.09582456201314926, -0.08239385485649109, -0.10547570884227753, -0.1369457095861435, 0.10373605042695999, 0.04170820116996765, -0.0013002310879528522, -0.10634888708591461, 0.1202116459608078, 0.0637342631816864, -0.05771591514348984, 0.03301641345024109, 0.04447147995233536, 0.0422622449696064, 0.02316172420978546, -0.05291995033621788, 0.11675518751144409, -0.0682515799999237, -0.1570422351360321, -0.059954941272735596, 0.08724265545606613, 0.06286325305700302, 0.03869328275322914, -0.0003074700362049043, 0.014051562175154686, 0.0038968920707702637, -0.09911571443080902, 0.04771164432168007, -0.05677508935332298, 0.04658115282654762, 0.019555145874619484, -0.04233933240175247, -0.04169389232993126, -0.05629970133304596, -0.019787924364209175, 0.10518011450767517, 0.3160676062107086, -0.07153476774692535, -0.03558922931551933, 0.05581295117735863, -0.04765762388706207, -0.16956070065498352, 0.13130375742912292, 0.04959776625037193, -0.0039269812405109406, 0.08978140354156494, -0.11217707395553589, 0.16616889834403992, 0.12772901356220245, -0.013660751283168793, 0.1259898692369461, -0.28816965222358704, -0.14091238379478455, 0.09083720296621323, 0.20136894285678864, 0.15352042019367218, -0.15650342404842377, -0.016253294423222542, -0.02784804441034794, -0.06211239844560623, 0.09764353185892105, -0.2131725251674652, 0.08326923847198486, -0.0029054980259388685, 0.06384819000959396, -0.004045946057885885, -0.05788479745388031, 0.12408086657524109, -0.006037900224328041, 0.16695380210876465, -0.04107246175408363, -0.020365092903375626, 0.07602081447839737, -0.01660791039466858, 0.005777337122708559, -0.05887968838214874, 0.022105561569333076, 0.0024067352060228586, -0.008980578742921352, -0.08979840576648712, 0.05877958983182907, -0.0444951094686985, -0.05832414701581001, -0.04680711403489113, 0.026822008192539215, -0.011301501654088497, -0.029352113604545593, 0.077999547123909, 0.023345978930592537, 0.18966694176197052, 0.06957823783159256, 0.04559558257460594, -0.0887220948934555, -0.020970633253455162, 0.0317574143409729, -0.02680680714547634, 0.06325152516365051, -0.1483703851699829, 0.015788480639457703, 0.11357226967811584, 0.021141160279512405, 0.1058540865778923, 0.07865608483552933, -0.05906281992793083, 0.028905466198921204, 0.08836241066455841, -0.15908509492874146, -0.10749784111976624, 0.019674627110362053, -0.05204217880964279, -0.07998134940862656, 0.09652642905712128, 0.08436348289251328, -0.07830183953046799, -0.00036837777588516474, -0.03200670704245567, -0.01188872754573822, -0.07204770296812057, 0.2164989858865738, 0.09793961048126221, 0.036136843264102936, -0.09819954633712769, 0.07309725135564804, 0.02865435555577278, -0.04351402819156647, 0.0033437968231737614, 0.06471860408782959, -0.05210549384355545, -0.0077495635487139225, 0.1388043761253357, 0.27215638756752014, -0.03365563228726387, -0.05459513142704964, -0.1766616702079773, -0.11601465940475464, 0.028953399509191513, 0.2169502228498459, 0.0954296663403511, -0.03151430934667587, 0.010932178236544132, 0.0299333818256855, -0.14162983000278473, 0.0791403129696846, 0.030834300443530083, 0.08008871972560883, -0.1423075944185257, 0.16555646061897278, 0.021278852596879005, 0.007647642865777016, -0.03211507946252823, 0.07043670117855072, -0.13582885265350342, 0.01632077246904373, -0.13201415538787842, -0.05632263794541359, 0.016223864629864693, -0.01827503740787506, 0.010423079133033752, -0.07603051513433456, -0.09197057783603668, 0.018874118104577065, -0.1199226826429367, -0.002940228907391429, 0.06942407041788101, 0.02731010876595974, -0.1588313728570938, -0.03269410505890846, 0.013069729320704937, -0.04232713580131531, 0.026398412883281708, 0.03563773259520531, 0.03441454470157623, 0.09330826997756958, -0.23648421466350555, -0.007821934297680855, 0.08505818247795105, -0.012864555232226849, 0.09176917374134064, -0.03594166785478592, -0.020347528159618378, -0.002296276856213808, 0.12170306593179703, 0.02142483927309513, 0.07121764123439789, -0.11806105822324753, 0.012077946215867996, -0.042279649525880814, -0.08308293670415878, -0.04039546474814415, -0.00035275635309517384, 0.07781483978033066, 0.0022578164935112, 0.1839645653963089, -0.10458476096391678, 0.022089118137955666, -0.2441759556531906, -0.014065814204514027, -0.013699964620172977, -0.07922256737947464, -0.09746915847063065, -0.030482230708003044, 0.08048564195632935, -0.05667272210121155, 0.14231649041175842, 0.037857819348573685, 0.05044537037611008, 0.03994894400238991, -0.04124104604125023, 0.0019092705333605409, 0.03870560601353645, 0.19938331842422485, 0.02399257756769657, -0.01579473353922367, 0.033772971481084824, 0.07078048586845398, 0.12007778137922287, 0.04401105269789696, 0.2436465620994568, 0.1845739781856537, -0.05584459751844406, 0.09448985755443573, 0.06797622889280319, -0.07598133385181427, -0.08744663000106812, 0.04954787716269493, -0.07196015864610672, 0.044650644063949585, -0.029583685100078583, 0.18089930713176727, 0.10563275963068008, -0.1629541665315628, 0.025733955204486847, -0.06109458953142166, -0.07988164573907852, -0.10174458473920822, 0.037627771496772766, -0.08944647759199142, -0.176961749792099, 0.030112434178590775, -0.10369785875082016, -0.003339817514643073, 0.11083023250102997, -0.0005041859112679958, -0.01084650307893753, 0.22599326074123383, 0.06536911427974701, 0.06668571382761002, 0.04099439084529877, 0.005007007159292698, -0.0409887433052063, -0.07379558682441711, -0.07288259267807007, -0.00877245981246233, -0.03393476456403732, 0.014990019612014294, -0.06241437420248985, -0.10070294141769409, 0.0527486652135849, 0.006128411740064621, -0.10147220641374588, 0.025314072147011757, 0.017552819103002548, 0.059528350830078125, -0.000827298907097429, 0.01783302240073681, 0.022000011056661606, -0.020677490159869194, 0.21140992641448975, -0.06483927369117737, -0.07985509932041168, -0.08008046448230743, 0.232949897646904, 0.020912842825055122, 0.006814600434154272, -0.00013276786194182932, -0.10278342664241791, 0.01992342248558998, 0.18827509880065918, 0.1785396933555603, -0.12445025146007538, 0.00477113900706172, -0.05005427449941635, -0.019337307661771774, -0.08877424895763397, 0.13135893642902374, 0.11908654123544693, 0.03284610062837601, -0.10072793811559677, -0.037411440163850784, -0.05653712898492813, -0.005864036735147238, -0.044109780341386795, 0.03639160841703415, 0.058109406381845474, 0.030181262642145157, -0.06097199767827988, 0.08336257189512253, -0.003390889149159193, -0.1376967579126358, 0.07729364186525345, -0.18727301061153412, -0.1539873629808426, -0.014404174871742725, 0.13067248463630676, -0.02252996154129505, 0.07286541908979416, -0.051656756550073624, -0.0017583409789949656, 0.022276069968938828, -0.04082920029759407, -0.05628272518515587, -0.1277117282152176, 0.07841237634420395, -0.15027748048305511, 0.23638257384300232, -0.03852364420890808, 0.09654883295297623, 0.11487561464309692, 0.024197401478886604, -0.060399871319532394, 0.09128851443529129, 0.031506385654211044, -0.11625425517559052, -0.001986507326364517, 0.09509388357400894, -0.04125906154513359, 0.060774337500333786, 0.04994227737188339, -0.1247624084353447, 0.026833783835172653, -0.05387754365801811, -0.06881856173276901, -0.06299179792404175, -0.046479880809783936, -0.0718105286359787, 0.10033164918422699, 0.17869895696640015, -0.014474369585514069, 0.06498916447162628, -0.06763304769992828, 0.05661049485206604, 0.07332714647054672, 0.05824404954910278, -0.04946725815534592, -0.26483577489852905, 0.055365659296512604, 0.13158737123012543, -0.052658114582300186, -0.24313019216060638, -0.0785665512084961, 0.009605697356164455, -0.07349766790866852, -0.08202303200960159, 0.06165676191449165, 0.11828239262104034, 0.0673544853925705, -0.05193069204688072, -0.15363138914108276, -0.0833745077252388, 0.16149716079235077, -0.1268071085214615, -0.09888780117034912 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text-generation
aidonuts/marigold-007-ep3
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T15:58:45+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 60, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.04654794931411743, 0.16618601977825165, -0.005445904564112425, 0.01853804849088192, 0.0981811136007309, 0.011998992413282394, 0.06433123350143433, 0.11398410052061081, -0.0230073444545269, 0.11406639218330383, 0.03047988750040531, 0.10172267258167267, 0.11317981779575348, 0.14841650426387787, -0.002152352826669812, -0.22403094172477722, 0.050844956189394, -0.12105348706245422, -0.033293843269348145, 0.11749980598688126, 0.1483822613954544, -0.09928343445062637, 0.07274559140205383, -0.029687678441405296, -0.012143402360379696, -0.030057786032557487, -0.05890674889087677, -0.046214159578084946, 0.04651786759495735, 0.06640566885471344, 0.06770290434360504, 0.0071083661168813705, 0.09012923389673233, -0.2696533799171448, 0.018959321081638336, 0.07145345956087112, -0.002759667346253991, 0.06957992166280746, 0.06404146552085876, -0.07107418030500412, 0.10337356477975845, -0.05106033384799957, 0.14650006592273712, 0.08365883678197861, -0.09081148356199265, -0.1895141303539276, -0.08866965025663376, 0.09882009029388428, 0.17572562396526337, 0.04925641790032387, -0.02320658043026924, 0.09761467576026917, -0.08769196271896362, 0.015438909642398357, 0.04981724172830582, -0.07620415836572647, -0.05378096550703049, 0.05986575037240982, 0.07907199114561081, 0.06627275794744492, -0.12434766441583633, -0.02885502204298973, 0.005009706597775221, 0.010980482213199139, 0.0769270583987236, 0.01728810742497444, 0.146672785282135, 0.0338633768260479, -0.12615777552127838, -0.04880760237574577, 0.09869225323200226, 0.03395522013306618, -0.04422314465045929, -0.24749068915843964, -0.03152675926685333, -0.030810698866844177, -0.029386121779680252, -0.03716538846492767, 0.04340358078479767, -0.007673026993870735, 0.08638741075992584, -0.0060646249912679195, -0.07403432577848434, -0.03937075287103653, 0.06169692054390907, 0.0672287791967392, 0.02999979443848133, -0.013745363801717758, 0.010938193649053574, 0.11620724946260452, 0.1095694974064827, -0.12054188549518585, -0.05555335059762001, -0.06393084675073624, -0.08656639605760574, -0.040790557861328125, 0.034162238240242004, 0.03456587344408035, 0.05349370837211609, 0.25305667519569397, 0.015654386952519417, 0.059652652591466904, 0.034477248787879944, 0.007892133668065071, 0.05848940089344978, 0.11044429242610931, -0.06018859148025513, -0.10444226115942001, -0.02648012898862362, 0.08843598514795303, 0.008199662901461124, -0.03287925571203232, -0.05088530853390694, 0.06019928678870201, 0.01946467161178589, 0.11926145106554031, 0.09061790257692337, 0.010536285117268562, -0.07121123373508453, -0.061038948595523834, 0.1891259253025055, -0.16544590890407562, 0.04322727024555206, 0.035097137093544006, -0.03903156518936157, 0.00019933005387429148, 0.013914269395172596, 0.016625655815005302, -0.025983380153775215, 0.09017423540353775, -0.054113563150167465, -0.04145489260554314, -0.11186197400093079, -0.03383193537592888, 0.033762916922569275, 0.008953776210546494, -0.035059962421655655, -0.033713940531015396, -0.08351044356822968, -0.07577689737081528, 0.09320491552352905, -0.07346344739198685, -0.04878907650709152, -0.01804324984550476, -0.07530532777309418, 0.022395428270101547, 0.019394835457205772, 0.07707412540912628, -0.02362251654267311, 0.04399976506829262, -0.05189276114106178, 0.05863580107688904, 0.11207318305969238, 0.03570080175995827, -0.05736649036407471, 0.06062258034944534, -0.23834340274333954, 0.09552820026874542, -0.07409077137708664, 0.05591456592082977, -0.153293639421463, -0.024439791217446327, 0.04788333550095558, 0.008784620091319084, -0.009650949388742447, 0.13416339457035065, -0.21702027320861816, -0.02536402828991413, 0.1717337965965271, -0.10057014971971512, -0.07069246470928192, 0.05619903281331062, -0.04835370555520058, 0.10988964140415192, 0.03825836628675461, -0.025690359994769096, 0.06171267107129097, -0.1267417073249817, 0.003717758459970355, -0.05005312338471413, -0.017048977315425873, 0.1548657864332199, 0.07182947546243668, -0.07217690348625183, 0.07399354875087738, 0.025708531960844994, -0.0246540866792202, -0.04625825211405754, -0.015164627693593502, -0.10536660254001617, 0.014689887873828411, -0.06369215250015259, 0.014470234513282776, -0.020807426422834396, -0.09071163833141327, -0.027962757274508476, -0.17504668235778809, -0.03014434315264225, 0.08651752024888992, -0.008693269453942776, -0.01803150773048401, -0.1178668737411499, 0.009341353550553322, 0.04177580401301384, 0.0061247628182172775, -0.13462838530540466, -0.04812471568584442, 0.02780051715672016, -0.1600649207830429, 0.034652888774871826, -0.05392369255423546, 0.04932025074958801, 0.025790516287088394, -0.028889117762446404, -0.026493212208151817, 0.021633783355355263, 0.005992184858769178, -0.011999987065792084, -0.24343903362751007, -0.028118690475821495, -0.024888472631573677, 0.1682123839855194, -0.20917098224163055, 0.03546025976538658, 0.07867541164159775, 0.15366052091121674, 0.011240328662097454, -0.04177491366863251, 0.005974748637527227, -0.06935794651508331, -0.02736494317650795, -0.05875484645366669, -0.0047869328409433365, -0.03310677409172058, -0.04545191675424576, 0.04568447172641754, -0.16510973870754242, -0.032636504620313644, 0.09776268899440765, 0.06289951503276825, -0.13922683894634247, -0.020621931180357933, -0.03630133345723152, -0.049253206700086594, -0.04911839962005615, -0.0605199858546257, 0.10893940925598145, 0.05891856551170349, 0.04574795812368393, -0.05928509309887886, -0.07568105310201645, -0.001827909960411489, -0.013898161239922047, -0.017864689230918884, 0.09759635478258133, 0.0751434788107872, -0.13251115381717682, 0.09224759042263031, 0.09603385627269745, 0.07919023185968399, 0.09113933145999908, -0.02355697751045227, -0.08261934667825699, -0.045987509191036224, 0.031442027539014816, 0.020124373957514763, 0.13039541244506836, -0.024294709786772728, 0.04352088272571564, 0.042134687304496765, -0.019369594752788544, 0.014752166345715523, -0.08687400817871094, 0.033972494304180145, 0.028472330421209335, -0.016721390187740326, 0.050190530717372894, -0.03876714035868645, 0.02440318465232849, 0.08830609917640686, 0.045322712510824203, 0.03507532551884651, 0.015493292361497879, -0.05206458270549774, -0.1083620935678482, 0.16405931115150452, -0.12714070081710815, -0.22483378648757935, -0.13936103880405426, 0.0037376401014626026, 0.035628627985715866, -0.015835661441087723, 0.002417160663753748, -0.059374887496232986, -0.12220635265111923, -0.08858037739992142, 0.015140829607844353, 0.04942670464515686, -0.09028962254524231, -0.06437795609235764, 0.058117836713790894, 0.03889724239706993, -0.14560972154140472, 0.017612040042877197, 0.04854894429445267, -0.09789852797985077, -0.006774199660867453, 0.08094939589500427, 0.0698540136218071, 0.1770169734954834, 0.017703235149383545, -0.021850809454917908, 0.032354529947042465, 0.20614571869373322, -0.13538233935832977, 0.11083246022462845, 0.13607586920261383, -0.09041404724121094, 0.08072979003190994, 0.19951270520687103, 0.03932560607790947, -0.10153959691524506, 0.031980328261852264, 0.02283124253153801, -0.0284719280898571, -0.24526868760585785, -0.07212468236684799, -0.004402178805321455, -0.058010730892419815, 0.07660572230815887, 0.09286724030971527, 0.08215958625078201, 0.012304253876209259, -0.09310996532440186, -0.08154371380805969, 0.05942574888467789, 0.10367169976234436, 0.024584239348769188, -0.010839897207915783, 0.08998730033636093, -0.034100502729415894, 0.019626356661319733, 0.0853661298751831, 0.005239574704319239, 0.17840281128883362, 0.05159219726920128, 0.18830420076847076, 0.07925192266702652, 0.07219027727842331, 0.009912233799695969, 0.013080619275569916, 0.018877580761909485, 0.03300119563937187, -0.002769160782918334, -0.08440786600112915, -0.02248465269804001, 0.11566436290740967, 0.06668911874294281, 0.010815348476171494, 0.015172341838479042, -0.04104290530085564, 0.07965951412916183, 0.1831512451171875, -0.007656289264559746, -0.1783534437417984, -0.057547420263290405, 0.07553383708000183, -0.09879875183105469, -0.09854305535554886, -0.013454320840537548, 0.03072015568614006, -0.17046253383159637, 0.023390959948301315, -0.02239842526614666, 0.1106182336807251, -0.14194999635219574, -0.020490378141403198, 0.07218493521213531, 0.07199500501155853, 0.004729843698441982, 0.05758659541606903, -0.16417601704597473, 0.10671813786029816, 0.008950476534664631, 0.06779605895280838, -0.09610627591609955, 0.1008887067437172, -0.004196076653897762, -0.02063460275530815, 0.1393408179283142, 0.002700034761801362, -0.06884108483791351, -0.0763031542301178, -0.08754398673772812, -0.009632662869989872, 0.12754282355308533, -0.1419651061296463, 0.08767123520374298, -0.037212442606687546, -0.0424150750041008, -0.0017086371080949903, -0.10206665843725204, -0.11638247221708298, -0.18888559937477112, 0.06001543253660202, -0.13492922484874725, 0.03152317553758621, -0.10799519717693329, -0.032371897250413895, -0.030304040759801865, 0.19337286055088043, -0.23447458446025848, -0.07199826091527939, -0.1475764364004135, -0.10233612358570099, 0.1443224400281906, -0.0501345656812191, 0.08485390990972519, -0.007241467013955116, 0.16846685111522675, 0.019060896709561348, -0.02531743235886097, 0.0971490666270256, -0.09173708409070969, -0.19302815198898315, -0.07869284600019455, 0.15662524104118347, 0.13260218501091003, 0.031680017709732056, -0.002461588243022561, 0.036563750356435776, -0.015421539545059204, -0.11935004591941833, 0.015969349071383476, 0.1787186712026596, 0.06237189099192619, 0.02331034652888775, -0.027346095070242882, -0.11273157596588135, -0.06900003552436829, -0.028530338779091835, 0.03054865077137947, 0.17762407660484314, -0.07057618349790573, 0.18207968771457672, 0.14163152873516083, -0.05922834202647209, -0.20400173962116241, 0.010538800619542599, 0.03055560030043125, 0.0009220078936778009, 0.02591954916715622, -0.20123432576656342, 0.08688826113939285, 0.004683020059019327, -0.05110127478837967, 0.13194532692432404, -0.17217805981636047, -0.14451217651367188, 0.0765485092997551, 0.038384392857551575, -0.19559739530086517, -0.12913893163204193, -0.09174312651157379, -0.045869920402765274, -0.18591414391994476, 0.09569250047206879, 0.0305706188082695, 0.010893458500504494, 0.03030681423842907, 0.029179483652114868, 0.019487828016281128, -0.0418255440890789, 0.18391458690166473, -0.024792250245809555, 0.026594700291752815, -0.08539514988660812, -0.06927408277988434, 0.03743394836783409, -0.052842434495687485, 0.07349982857704163, -0.023486759513616562, 0.007861839607357979, -0.10348054021596909, -0.042148489505052567, -0.03735732287168503, 0.015448716469109058, -0.09657872468233109, -0.08514349907636642, -0.045032672584056854, 0.09675803780555725, 0.09690850973129272, -0.033646680414676666, -0.028050623834133148, -0.07533035427331924, 0.04412057250738144, 0.19926515221595764, 0.1785389482975006, 0.042153384536504745, -0.08034496754407883, -0.004150947090238333, -0.010121207684278488, 0.04310847446322441, -0.20463712513446808, 0.06283636391162872, 0.05450061708688736, 0.01973269321024418, 0.11436162889003754, -0.019565396010875702, -0.15359151363372803, -0.07263088971376419, 0.06303015351295471, -0.060181066393852234, -0.19620554149150848, 0.00867035984992981, 0.060603946447372437, -0.16371412575244904, -0.04535605385899544, 0.04643881320953369, -0.005620351992547512, -0.038163937628269196, 0.021896906197071075, 0.09194854646921158, 0.0026654244866222143, 0.07427921891212463, 0.05387866869568825, 0.0827430784702301, -0.10537070035934448, 0.08090532571077347, 0.08839722722768784, -0.08452684432268143, 0.023530138656497, 0.10478579998016357, -0.059433579444885254, -0.03440561518073082, 0.020135708153247833, 0.08153781294822693, 0.01775863952934742, -0.040019966661930084, 0.013229827396571636, -0.10452935844659805, 0.05954122915863991, 0.08839859813451767, 0.032507482916116714, 0.016702456399798393, 0.03425082191824913, 0.04607953503727913, -0.07238735258579254, 0.12142276018857956, 0.031868141144514084, 0.017129309475421906, -0.036505792289972305, -0.040896978229284286, 0.019542274996638298, -0.03214648738503456, -0.005015232600271702, -0.03023446537554264, -0.07695909589529037, -0.014793801121413708, -0.1626158058643341, -0.011131818406283855, -0.05648450180888176, 0.010329355485737324, 0.03204665705561638, -0.032609567046165466, 0.008124498650431633, 0.009250079281628132, -0.07695289701223373, -0.0663459524512291, -0.020460480824112892, 0.09540658444166183, -0.16213038563728333, 0.022481130436062813, 0.08244425803422928, -0.12187694013118744, 0.09281346201896667, 0.016204802319407463, -0.006236857734620571, 0.025038830935955048, -0.1475188434123993, 0.034843120723962784, -0.03386561945080757, 0.010836300440132618, 0.04373383894562721, -0.21569781005382538, -0.00004886732858722098, -0.033673107624053955, -0.06639216095209122, -0.009451326914131641, -0.03672455996274948, -0.11508306115865707, 0.1058407872915268, 0.007236586883664131, -0.08753558248281479, -0.03186136856675148, 0.029325377196073532, 0.0838974118232727, -0.021959776058793068, 0.15145497024059296, -0.008370938710868359, 0.07429654151201248, -0.16209737956523895, -0.018623165786266327, -0.006028574425727129, 0.022658247500658035, -0.01664556935429573, -0.01111356820911169, 0.044031109660863876, -0.022746501490473747, 0.17925859987735748, -0.030318550765514374, 0.02272745408117771, 0.06815794110298157, 0.019072026014328003, -0.030184008181095123, 0.10406795144081116, 0.04094860330224037, 0.02014910988509655, 0.018591465428471565, 0.003289656015112996, -0.04647882282733917, -0.03173251822590828, -0.19407226145267487, 0.07288651913404465, 0.15608493983745575, 0.09729263186454773, -0.016707008704543114, 0.07954329252243042, -0.10199416428804398, -0.1109243705868721, 0.12477338314056396, -0.04797708988189697, -0.002418199321255088, -0.07150927931070328, 0.13247236609458923, 0.1437523066997528, -0.1859612911939621, 0.07269313186407089, -0.0699717253446579, -0.04708027467131615, -0.10980689525604248, -0.19441905617713928, -0.05561789125204086, -0.049456022679805756, -0.016053348779678345, -0.04698808491230011, 0.07504211366176605, 0.054538097232580185, 0.006766852922737598, -0.0023397188633680344, 0.06506035476922989, -0.031050674617290497, -0.0037882844917476177, 0.032597362995147705, 0.06591679900884628, 0.012734474614262581, -0.030802709981799126, 0.016619903966784477, -0.013545602560043335, 0.045626189559698105, 0.06578011065721512, 0.04976864159107208, -0.02938537672162056, 0.014603170566260815, -0.038539156317710876, -0.10249634087085724, 0.043612558394670486, -0.024421939626336098, -0.0789753645658493, 0.15477414429187775, 0.023680059239268303, 0.007779473438858986, -0.020137663930654526, 0.23901568353176117, -0.0738423764705658, -0.0964353010058403, -0.14737580716609955, 0.10557299107313156, -0.038081806153059006, 0.05800395458936691, 0.04625935107469559, -0.10226529091596603, 0.018044332042336464, 0.1338089406490326, 0.16182038187980652, -0.039008259773254395, 0.020095856860280037, 0.031135575845837593, 0.00566398398950696, -0.03622615709900856, 0.04847532883286476, 0.06906453520059586, 0.16569648683071136, -0.04632584750652313, 0.09100406616926193, 0.0019041687482967973, -0.09579581767320633, -0.038361791521310806, 0.11069868505001068, -0.016052277758717537, 0.019335128366947174, -0.05818064883351326, 0.11742528527975082, -0.06386786699295044, -0.23783175647258759, 0.06453443318605423, -0.0684293657541275, -0.13765870034694672, -0.02378307841718197, 0.08207765966653824, -0.012955902144312859, 0.027587108314037323, 0.0730307325720787, -0.07240920513868332, 0.201939657330513, 0.03798431158065796, -0.05499868467450142, -0.055047210305929184, 0.0805421993136406, -0.10008571296930313, 0.2739645540714264, 0.01557221356779337, 0.04601577669382095, 0.10384146869182587, -0.009341772645711899, -0.13838784396648407, 0.019836371764540672, 0.09581108391284943, -0.10502193123102188, 0.04196618124842644, 0.19815568625926971, -0.0014755994779989123, 0.12389086186885834, 0.07657600939273834, -0.07551808655261993, 0.0478031262755394, -0.08054235577583313, -0.06760486960411072, -0.09260394424200058, 0.09703279286623001, -0.07772123068571091, 0.14251399040222168, 0.13876807689666748, -0.05074559152126312, 0.012724342755973339, -0.031311117112636566, 0.044293127954006195, -0.00010600237874314189, 0.10321761667728424, 0.004272161517292261, -0.1832672357559204, 0.024692710489034653, 0.005650998093187809, 0.10749758034944534, -0.16033467650413513, -0.09566054493188858, 0.042343202978372574, 0.003505636239424348, -0.0672195628285408, 0.1290110945701599, 0.05665452033281326, 0.04342988133430481, -0.03997718170285225, -0.03521440550684929, -0.0060732318088412285, 0.13561366498470306, -0.10713256150484085, 0.0009933578548952937 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text-generation
unalignment/weeeeee.0
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T16:02:46+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 60, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.04654794931411743, 0.16618601977825165, -0.005445904564112425, 0.01853804849088192, 0.0981811136007309, 0.011998992413282394, 0.06433123350143433, 0.11398410052061081, -0.0230073444545269, 0.11406639218330383, 0.03047988750040531, 0.10172267258167267, 0.11317981779575348, 0.14841650426387787, -0.002152352826669812, -0.22403094172477722, 0.050844956189394, -0.12105348706245422, -0.033293843269348145, 0.11749980598688126, 0.1483822613954544, -0.09928343445062637, 0.07274559140205383, -0.029687678441405296, -0.012143402360379696, -0.030057786032557487, -0.05890674889087677, -0.046214159578084946, 0.04651786759495735, 0.06640566885471344, 0.06770290434360504, 0.0071083661168813705, 0.09012923389673233, -0.2696533799171448, 0.018959321081638336, 0.07145345956087112, -0.002759667346253991, 0.06957992166280746, 0.06404146552085876, -0.07107418030500412, 0.10337356477975845, -0.05106033384799957, 0.14650006592273712, 0.08365883678197861, -0.09081148356199265, -0.1895141303539276, -0.08866965025663376, 0.09882009029388428, 0.17572562396526337, 0.04925641790032387, -0.02320658043026924, 0.09761467576026917, -0.08769196271896362, 0.015438909642398357, 0.04981724172830582, -0.07620415836572647, -0.05378096550703049, 0.05986575037240982, 0.07907199114561081, 0.06627275794744492, -0.12434766441583633, -0.02885502204298973, 0.005009706597775221, 0.010980482213199139, 0.0769270583987236, 0.01728810742497444, 0.146672785282135, 0.0338633768260479, -0.12615777552127838, -0.04880760237574577, 0.09869225323200226, 0.03395522013306618, -0.04422314465045929, -0.24749068915843964, -0.03152675926685333, -0.030810698866844177, -0.029386121779680252, -0.03716538846492767, 0.04340358078479767, -0.007673026993870735, 0.08638741075992584, -0.0060646249912679195, -0.07403432577848434, -0.03937075287103653, 0.06169692054390907, 0.0672287791967392, 0.02999979443848133, -0.013745363801717758, 0.010938193649053574, 0.11620724946260452, 0.1095694974064827, -0.12054188549518585, -0.05555335059762001, -0.06393084675073624, -0.08656639605760574, -0.040790557861328125, 0.034162238240242004, 0.03456587344408035, 0.05349370837211609, 0.25305667519569397, 0.015654386952519417, 0.059652652591466904, 0.034477248787879944, 0.007892133668065071, 0.05848940089344978, 0.11044429242610931, -0.06018859148025513, -0.10444226115942001, -0.02648012898862362, 0.08843598514795303, 0.008199662901461124, -0.03287925571203232, -0.05088530853390694, 0.06019928678870201, 0.01946467161178589, 0.11926145106554031, 0.09061790257692337, 0.010536285117268562, -0.07121123373508453, -0.061038948595523834, 0.1891259253025055, -0.16544590890407562, 0.04322727024555206, 0.035097137093544006, -0.03903156518936157, 0.00019933005387429148, 0.013914269395172596, 0.016625655815005302, -0.025983380153775215, 0.09017423540353775, -0.054113563150167465, -0.04145489260554314, -0.11186197400093079, -0.03383193537592888, 0.033762916922569275, 0.008953776210546494, -0.035059962421655655, -0.033713940531015396, -0.08351044356822968, -0.07577689737081528, 0.09320491552352905, -0.07346344739198685, -0.04878907650709152, -0.01804324984550476, -0.07530532777309418, 0.022395428270101547, 0.019394835457205772, 0.07707412540912628, -0.02362251654267311, 0.04399976506829262, -0.05189276114106178, 0.05863580107688904, 0.11207318305969238, 0.03570080175995827, -0.05736649036407471, 0.06062258034944534, -0.23834340274333954, 0.09552820026874542, -0.07409077137708664, 0.05591456592082977, -0.153293639421463, -0.024439791217446327, 0.04788333550095558, 0.008784620091319084, -0.009650949388742447, 0.13416339457035065, -0.21702027320861816, -0.02536402828991413, 0.1717337965965271, -0.10057014971971512, -0.07069246470928192, 0.05619903281331062, -0.04835370555520058, 0.10988964140415192, 0.03825836628675461, -0.025690359994769096, 0.06171267107129097, -0.1267417073249817, 0.003717758459970355, -0.05005312338471413, -0.017048977315425873, 0.1548657864332199, 0.07182947546243668, -0.07217690348625183, 0.07399354875087738, 0.025708531960844994, -0.0246540866792202, -0.04625825211405754, -0.015164627693593502, -0.10536660254001617, 0.014689887873828411, -0.06369215250015259, 0.014470234513282776, -0.020807426422834396, -0.09071163833141327, -0.027962757274508476, -0.17504668235778809, -0.03014434315264225, 0.08651752024888992, -0.008693269453942776, -0.01803150773048401, -0.1178668737411499, 0.009341353550553322, 0.04177580401301384, 0.0061247628182172775, -0.13462838530540466, -0.04812471568584442, 0.02780051715672016, -0.1600649207830429, 0.034652888774871826, -0.05392369255423546, 0.04932025074958801, 0.025790516287088394, -0.028889117762446404, -0.026493212208151817, 0.021633783355355263, 0.005992184858769178, -0.011999987065792084, -0.24343903362751007, -0.028118690475821495, -0.024888472631573677, 0.1682123839855194, -0.20917098224163055, 0.03546025976538658, 0.07867541164159775, 0.15366052091121674, 0.011240328662097454, -0.04177491366863251, 0.005974748637527227, -0.06935794651508331, -0.02736494317650795, -0.05875484645366669, -0.0047869328409433365, -0.03310677409172058, -0.04545191675424576, 0.04568447172641754, -0.16510973870754242, -0.032636504620313644, 0.09776268899440765, 0.06289951503276825, -0.13922683894634247, -0.020621931180357933, -0.03630133345723152, -0.049253206700086594, -0.04911839962005615, -0.0605199858546257, 0.10893940925598145, 0.05891856551170349, 0.04574795812368393, -0.05928509309887886, -0.07568105310201645, -0.001827909960411489, -0.013898161239922047, -0.017864689230918884, 0.09759635478258133, 0.0751434788107872, -0.13251115381717682, 0.09224759042263031, 0.09603385627269745, 0.07919023185968399, 0.09113933145999908, -0.02355697751045227, -0.08261934667825699, -0.045987509191036224, 0.031442027539014816, 0.020124373957514763, 0.13039541244506836, -0.024294709786772728, 0.04352088272571564, 0.042134687304496765, -0.019369594752788544, 0.014752166345715523, -0.08687400817871094, 0.033972494304180145, 0.028472330421209335, -0.016721390187740326, 0.050190530717372894, -0.03876714035868645, 0.02440318465232849, 0.08830609917640686, 0.045322712510824203, 0.03507532551884651, 0.015493292361497879, -0.05206458270549774, -0.1083620935678482, 0.16405931115150452, -0.12714070081710815, -0.22483378648757935, -0.13936103880405426, 0.0037376401014626026, 0.035628627985715866, -0.015835661441087723, 0.002417160663753748, -0.059374887496232986, -0.12220635265111923, -0.08858037739992142, 0.015140829607844353, 0.04942670464515686, -0.09028962254524231, -0.06437795609235764, 0.058117836713790894, 0.03889724239706993, -0.14560972154140472, 0.017612040042877197, 0.04854894429445267, -0.09789852797985077, -0.006774199660867453, 0.08094939589500427, 0.0698540136218071, 0.1770169734954834, 0.017703235149383545, -0.021850809454917908, 0.032354529947042465, 0.20614571869373322, -0.13538233935832977, 0.11083246022462845, 0.13607586920261383, -0.09041404724121094, 0.08072979003190994, 0.19951270520687103, 0.03932560607790947, -0.10153959691524506, 0.031980328261852264, 0.02283124253153801, -0.0284719280898571, -0.24526868760585785, -0.07212468236684799, -0.004402178805321455, -0.058010730892419815, 0.07660572230815887, 0.09286724030971527, 0.08215958625078201, 0.012304253876209259, -0.09310996532440186, -0.08154371380805969, 0.05942574888467789, 0.10367169976234436, 0.024584239348769188, -0.010839897207915783, 0.08998730033636093, -0.034100502729415894, 0.019626356661319733, 0.0853661298751831, 0.005239574704319239, 0.17840281128883362, 0.05159219726920128, 0.18830420076847076, 0.07925192266702652, 0.07219027727842331, 0.009912233799695969, 0.013080619275569916, 0.018877580761909485, 0.03300119563937187, -0.002769160782918334, -0.08440786600112915, -0.02248465269804001, 0.11566436290740967, 0.06668911874294281, 0.010815348476171494, 0.015172341838479042, -0.04104290530085564, 0.07965951412916183, 0.1831512451171875, -0.007656289264559746, -0.1783534437417984, -0.057547420263290405, 0.07553383708000183, -0.09879875183105469, -0.09854305535554886, -0.013454320840537548, 0.03072015568614006, -0.17046253383159637, 0.023390959948301315, -0.02239842526614666, 0.1106182336807251, -0.14194999635219574, -0.020490378141403198, 0.07218493521213531, 0.07199500501155853, 0.004729843698441982, 0.05758659541606903, -0.16417601704597473, 0.10671813786029816, 0.008950476534664631, 0.06779605895280838, -0.09610627591609955, 0.1008887067437172, -0.004196076653897762, -0.02063460275530815, 0.1393408179283142, 0.002700034761801362, -0.06884108483791351, -0.0763031542301178, -0.08754398673772812, -0.009632662869989872, 0.12754282355308533, -0.1419651061296463, 0.08767123520374298, -0.037212442606687546, -0.0424150750041008, -0.0017086371080949903, -0.10206665843725204, -0.11638247221708298, -0.18888559937477112, 0.06001543253660202, -0.13492922484874725, 0.03152317553758621, -0.10799519717693329, -0.032371897250413895, -0.030304040759801865, 0.19337286055088043, -0.23447458446025848, -0.07199826091527939, -0.1475764364004135, -0.10233612358570099, 0.1443224400281906, -0.0501345656812191, 0.08485390990972519, -0.007241467013955116, 0.16846685111522675, 0.019060896709561348, -0.02531743235886097, 0.0971490666270256, -0.09173708409070969, -0.19302815198898315, -0.07869284600019455, 0.15662524104118347, 0.13260218501091003, 0.031680017709732056, -0.002461588243022561, 0.036563750356435776, -0.015421539545059204, -0.11935004591941833, 0.015969349071383476, 0.1787186712026596, 0.06237189099192619, 0.02331034652888775, -0.027346095070242882, -0.11273157596588135, -0.06900003552436829, -0.028530338779091835, 0.03054865077137947, 0.17762407660484314, -0.07057618349790573, 0.18207968771457672, 0.14163152873516083, -0.05922834202647209, -0.20400173962116241, 0.010538800619542599, 0.03055560030043125, 0.0009220078936778009, 0.02591954916715622, -0.20123432576656342, 0.08688826113939285, 0.004683020059019327, -0.05110127478837967, 0.13194532692432404, -0.17217805981636047, -0.14451217651367188, 0.0765485092997551, 0.038384392857551575, -0.19559739530086517, -0.12913893163204193, -0.09174312651157379, -0.045869920402765274, -0.18591414391994476, 0.09569250047206879, 0.0305706188082695, 0.010893458500504494, 0.03030681423842907, 0.029179483652114868, 0.019487828016281128, -0.0418255440890789, 0.18391458690166473, -0.024792250245809555, 0.026594700291752815, -0.08539514988660812, -0.06927408277988434, 0.03743394836783409, -0.052842434495687485, 0.07349982857704163, -0.023486759513616562, 0.007861839607357979, -0.10348054021596909, -0.042148489505052567, -0.03735732287168503, 0.015448716469109058, -0.09657872468233109, -0.08514349907636642, -0.045032672584056854, 0.09675803780555725, 0.09690850973129272, -0.033646680414676666, -0.028050623834133148, -0.07533035427331924, 0.04412057250738144, 0.19926515221595764, 0.1785389482975006, 0.042153384536504745, -0.08034496754407883, -0.004150947090238333, -0.010121207684278488, 0.04310847446322441, -0.20463712513446808, 0.06283636391162872, 0.05450061708688736, 0.01973269321024418, 0.11436162889003754, -0.019565396010875702, -0.15359151363372803, -0.07263088971376419, 0.06303015351295471, -0.060181066393852234, -0.19620554149150848, 0.00867035984992981, 0.060603946447372437, -0.16371412575244904, -0.04535605385899544, 0.04643881320953369, -0.005620351992547512, -0.038163937628269196, 0.021896906197071075, 0.09194854646921158, 0.0026654244866222143, 0.07427921891212463, 0.05387866869568825, 0.0827430784702301, -0.10537070035934448, 0.08090532571077347, 0.08839722722768784, -0.08452684432268143, 0.023530138656497, 0.10478579998016357, -0.059433579444885254, -0.03440561518073082, 0.020135708153247833, 0.08153781294822693, 0.01775863952934742, -0.040019966661930084, 0.013229827396571636, -0.10452935844659805, 0.05954122915863991, 0.08839859813451767, 0.032507482916116714, 0.016702456399798393, 0.03425082191824913, 0.04607953503727913, -0.07238735258579254, 0.12142276018857956, 0.031868141144514084, 0.017129309475421906, -0.036505792289972305, -0.040896978229284286, 0.019542274996638298, -0.03214648738503456, -0.005015232600271702, -0.03023446537554264, -0.07695909589529037, -0.014793801121413708, -0.1626158058643341, -0.011131818406283855, -0.05648450180888176, 0.010329355485737324, 0.03204665705561638, -0.032609567046165466, 0.008124498650431633, 0.009250079281628132, -0.07695289701223373, -0.0663459524512291, -0.020460480824112892, 0.09540658444166183, -0.16213038563728333, 0.022481130436062813, 0.08244425803422928, -0.12187694013118744, 0.09281346201896667, 0.016204802319407463, -0.006236857734620571, 0.025038830935955048, -0.1475188434123993, 0.034843120723962784, -0.03386561945080757, 0.010836300440132618, 0.04373383894562721, -0.21569781005382538, -0.00004886732858722098, -0.033673107624053955, -0.06639216095209122, -0.009451326914131641, -0.03672455996274948, -0.11508306115865707, 0.1058407872915268, 0.007236586883664131, -0.08753558248281479, -0.03186136856675148, 0.029325377196073532, 0.0838974118232727, -0.021959776058793068, 0.15145497024059296, -0.008370938710868359, 0.07429654151201248, -0.16209737956523895, -0.018623165786266327, -0.006028574425727129, 0.022658247500658035, -0.01664556935429573, -0.01111356820911169, 0.044031109660863876, -0.022746501490473747, 0.17925859987735748, -0.030318550765514374, 0.02272745408117771, 0.06815794110298157, 0.019072026014328003, -0.030184008181095123, 0.10406795144081116, 0.04094860330224037, 0.02014910988509655, 0.018591465428471565, 0.003289656015112996, -0.04647882282733917, -0.03173251822590828, -0.19407226145267487, 0.07288651913404465, 0.15608493983745575, 0.09729263186454773, -0.016707008704543114, 0.07954329252243042, -0.10199416428804398, -0.1109243705868721, 0.12477338314056396, -0.04797708988189697, -0.002418199321255088, -0.07150927931070328, 0.13247236609458923, 0.1437523066997528, -0.1859612911939621, 0.07269313186407089, -0.0699717253446579, -0.04708027467131615, -0.10980689525604248, -0.19441905617713928, -0.05561789125204086, -0.049456022679805756, -0.016053348779678345, -0.04698808491230011, 0.07504211366176605, 0.054538097232580185, 0.006766852922737598, -0.0023397188633680344, 0.06506035476922989, -0.031050674617290497, -0.0037882844917476177, 0.032597362995147705, 0.06591679900884628, 0.012734474614262581, -0.030802709981799126, 0.016619903966784477, -0.013545602560043335, 0.045626189559698105, 0.06578011065721512, 0.04976864159107208, -0.02938537672162056, 0.014603170566260815, -0.038539156317710876, -0.10249634087085724, 0.043612558394670486, -0.024421939626336098, -0.0789753645658493, 0.15477414429187775, 0.023680059239268303, 0.007779473438858986, -0.020137663930654526, 0.23901568353176117, -0.0738423764705658, -0.0964353010058403, -0.14737580716609955, 0.10557299107313156, -0.038081806153059006, 0.05800395458936691, 0.04625935107469559, -0.10226529091596603, 0.018044332042336464, 0.1338089406490326, 0.16182038187980652, -0.039008259773254395, 0.020095856860280037, 0.031135575845837593, 0.00566398398950696, -0.03622615709900856, 0.04847532883286476, 0.06906453520059586, 0.16569648683071136, -0.04632584750652313, 0.09100406616926193, 0.0019041687482967973, -0.09579581767320633, -0.038361791521310806, 0.11069868505001068, -0.016052277758717537, 0.019335128366947174, -0.05818064883351326, 0.11742528527975082, -0.06386786699295044, -0.23783175647258759, 0.06453443318605423, -0.0684293657541275, -0.13765870034694672, -0.02378307841718197, 0.08207765966653824, -0.012955902144312859, 0.027587108314037323, 0.0730307325720787, -0.07240920513868332, 0.201939657330513, 0.03798431158065796, -0.05499868467450142, -0.055047210305929184, 0.0805421993136406, -0.10008571296930313, 0.2739645540714264, 0.01557221356779337, 0.04601577669382095, 0.10384146869182587, -0.009341772645711899, -0.13838784396648407, 0.019836371764540672, 0.09581108391284943, -0.10502193123102188, 0.04196618124842644, 0.19815568625926971, -0.0014755994779989123, 0.12389086186885834, 0.07657600939273834, -0.07551808655261993, 0.0478031262755394, -0.08054235577583313, -0.06760486960411072, -0.09260394424200058, 0.09703279286623001, -0.07772123068571091, 0.14251399040222168, 0.13876807689666748, -0.05074559152126312, 0.012724342755973339, -0.031311117112636566, 0.044293127954006195, -0.00010600237874314189, 0.10321761667728424, 0.004272161517292261, -0.1832672357559204, 0.024692710489034653, 0.005650998093187809, 0.10749758034944534, -0.16033467650413513, -0.09566054493188858, 0.042343202978372574, 0.003505636239424348, -0.0672195628285408, 0.1290110945701599, 0.05665452033281326, 0.04342988133430481, -0.03997718170285225, -0.03521440550684929, -0.0060732318088412285, 0.13561366498470306, -0.10713256150484085, 0.0009933578548952937 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilroberta-base-finetuned-wikitext2 This model is a fine-tuned version of [distilroberta-base](https://huggingface.co/distilroberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 2.9870 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 2e-05 - train_batch_size: 16 - eval_batch_size: 16 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 2 ### Training results | Training Loss | Epoch | Step | Validation Loss | |:-------------:|:-----:|:----:|:---------------:| | 3.3087 | 1.0 | 4001 | 3.1219 | | 3.1437 | 2.0 | 8002 | 2.9870 | ### Framework versions - Transformers 4.37.0 - Pytorch 2.1.2 - Datasets 2.1.0 - Tokenizers 0.15.1
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "base_model": "distilroberta-base", "model-index": [{"name": "distilroberta-base-finetuned-wikitext2", "results": []}]}
fill-mask
MagdyNasr/distilroberta-base-finetuned-wikitext2
[ "transformers", "tensorboard", "safetensors", "roberta", "fill-mask", "generated_from_trainer", "base_model:distilroberta-base", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-10T16:05:54+00:00
[]
[]
TAGS #transformers #tensorboard #safetensors #roberta #fill-mask #generated_from_trainer #base_model-distilroberta-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us
distilroberta-base-finetuned-wikitext2 ====================================== This model is a fine-tuned version of distilroberta-base on the None dataset. It achieves the following results on the evaluation set: * Loss: 2.9870 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 2e-05 * train\_batch\_size: 16 * eval\_batch\_size: 16 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 2 ### Training results ### Framework versions * Transformers 4.37.0 * Pytorch 2.1.2 * Datasets 2.1.0 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.1.0\n* Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #tensorboard #safetensors #roberta #fill-mask #generated_from_trainer #base_model-distilroberta-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.1.0\n* Tokenizers 0.15.1" ]
[ 68, 98, 4, 30 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #roberta #fill-mask #generated_from_trainer #base_model-distilroberta-base #license-apache-2.0 #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 2e-05\n* train\\_batch\\_size: 16\n* eval\\_batch\\_size: 16\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 2### Training results### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.1.0\n* Tokenizers 0.15.1" ]
[ -0.08627286553382874, 0.07090317457914352, -0.0027419745456427336, 0.10829684883356094, 0.13170917332172394, 0.022251153364777565, 0.15960565209388733, 0.11391827464103699, -0.07898696511983871, 0.040176957845687866, 0.13823385536670685, 0.12857568264007568, 0.00971664022654295, 0.1581239104270935, -0.04375557228922844, -0.22989504039287567, 0.007963055744767189, 0.02695103920996189, -0.10133980959653854, 0.11057489365339279, 0.1024993285536766, -0.11917927116155624, 0.0833168476819992, 0.009324719198048115, -0.17620234191417694, 0.02320067025721073, 0.027421724051237106, -0.05919714644551277, 0.1289401352405548, 0.02778642065823078, 0.14149066805839539, 0.006406453438103199, 0.09669096022844315, -0.18227903544902802, 0.015242507681250572, 0.06278342753648758, -0.007511621806770563, 0.07528039067983627, 0.014062987640500069, 0.01095359493046999, 0.10042325407266617, -0.08968880027532578, 0.06269004195928574, 0.01733562909066677, -0.12098520249128342, -0.23936927318572998, -0.07852740585803986, 0.03413446247577667, 0.09353189915418625, 0.08721019327640533, -0.004600008483976126, 0.1448620706796646, -0.0432850643992424, 0.09227905422449112, 0.22574400901794434, -0.2979075610637665, -0.07102490216493607, 0.048199206590652466, 0.046285416930913925, 0.04575774818658829, -0.09500990808010101, -0.01370355673134327, 0.06425806134939194, 0.02961248904466629, 0.14692571759223938, -0.033178795129060745, -0.053794167935848236, -0.00024192330602090806, -0.14028248190879822, -0.024467092007398605, 0.15085092186927795, 0.049923013895750046, -0.047476932406425476, -0.04159298166632652, -0.07129627466201782, -0.12492474168539047, -0.039178233593702316, -0.028226103633642197, 0.048147477209568024, -0.028370054438710213, -0.0789402425289154, -0.02777259238064289, -0.10768239200115204, -0.08147196471691132, -0.060455322265625, 0.17376884818077087, 0.03770657628774643, 0.016475258395075798, -0.024552661925554276, 0.08268392086029053, -0.05828108638525009, -0.1373441219329834, 0.013148372992873192, 0.02910486049950123, 0.005028106737881899, -0.041192952543497086, -0.06131161004304886, -0.10051095485687256, 0.030316412448883057, 0.17402735352516174, -0.050665780901908875, 0.04123774915933609, 0.03565472736954689, 0.05315670743584633, -0.10336019843816757, 0.15844054520130157, -0.04172220453619957, -0.024731246754527092, 0.019660821184515953, 0.05748859420418739, 0.061409298330545425, -0.0015455491375178099, -0.11607003211975098, 0.026627078652381897, 0.07469645887613297, 0.004756190814077854, -0.05294761061668396, 0.06585802137851715, -0.050621043890714645, 0.00603399146348238, 0.008802722208201885, -0.08955729752779007, 0.021601693704724312, -0.020646680146455765, -0.049315959215164185, -0.07026137411594391, 0.04974799230694771, 0.021225864067673683, 0.02267792448401451, 0.10587070882320404, -0.09232598543167114, 0.003957690205425024, -0.09733257442712784, -0.11653555184602737, 0.015406539663672447, -0.08202114701271057, 0.03400713577866554, -0.10906919091939926, -0.19552408158779144, 0.009357892908155918, 0.08294185996055603, -0.03178269416093826, -0.031297795474529266, -0.02742084488272667, -0.07965173572301865, 0.016402030363678932, -0.008231507614254951, 0.09959220141172409, -0.06071244180202484, 0.09112851321697235, 0.056990090757608414, 0.08860330283641815, -0.04720207303762436, 0.028965307399630547, -0.09335552155971527, 0.03417288511991501, -0.19694073498249054, -0.0014084643917158246, -0.06487094610929489, 0.05999612435698509, -0.08010890334844589, -0.08087360858917236, -0.02514708787202835, 0.014314154163002968, 0.07749434560537338, 0.08096712827682495, -0.15970464050769806, -0.07025022059679031, 0.18135994672775269, -0.09661009907722473, -0.1302478313446045, 0.13110794126987457, -0.05441771820187569, 0.06209754943847656, 0.03910553455352783, 0.15883754193782806, 0.05433626472949982, -0.13176067173480988, 0.011144139803946018, -0.00033538744901306927, 0.033836450427770615, -0.044772062450647354, 0.07521583139896393, 0.0025379504077136517, 0.019953526556491852, 0.010687258094549179, -0.03764013573527336, 0.05626026540994644, -0.08085732161998749, -0.09341681003570557, -0.046625491231679916, -0.10821249336004257, 0.05498361214995384, 0.051801297813653946, 0.06354965269565582, -0.11695011705160141, -0.10840702056884766, 0.04238699749112129, 0.07921275496482849, -0.06435087323188782, 0.020898953080177307, -0.07696326076984406, 0.07516317814588547, -0.07055049389600754, -0.023810427635908127, -0.14515723288059235, -0.05117565393447876, 0.005495747551321983, -0.02653207816183567, 0.017611617222428322, 0.007030869368463755, 0.08772122114896774, 0.08351001143455505, -0.07315938174724579, -0.02090776525437832, -0.046704791486263275, 0.014979507774114609, -0.10687322169542313, -0.20536565780639648, -0.02696146070957184, -0.04343528673052788, 0.12474467605352402, -0.19715166091918945, 0.04511271417140961, -0.04596549645066261, 0.09405029565095901, 0.04056459292769432, -0.010615374892950058, -0.049863677471876144, 0.05669483169913292, -0.03664024919271469, -0.06449770927429199, 0.042138200253248215, 0.008903494104743004, -0.09002410620450974, -0.04591687023639679, -0.12907245755195618, 0.178187757730484, 0.12637099623680115, -0.08522894233465195, -0.08993806689977646, 0.029729831963777542, -0.05898960679769516, -0.032770488411188126, -0.04796623811125755, 0.0007012076093815267, 0.11266376823186874, -0.022167330607771873, 0.13475604355335236, -0.07421935349702835, -0.03318658098578453, 0.0361710861325264, -0.06610476970672607, -0.001535702496767044, 0.07733708620071411, 0.11345460265874863, -0.08876517415046692, 0.1485835313796997, 0.18363358080387115, -0.11982356756925583, 0.14156870543956757, -0.03467357158660889, -0.06886133551597595, -0.034567367285490036, -0.005533203016966581, 0.014092579483985901, 0.13793036341667175, -0.11647256463766098, 0.012295492924749851, 0.015137108974158764, 0.0010212494526058435, 0.012601716443896294, -0.21818484365940094, -0.036760084331035614, 0.04109066724777222, -0.04382113367319107, -0.0011772826546803117, -0.002848472911864519, -0.02070111408829689, 0.0873776525259018, -0.001516292686574161, -0.08696916699409485, 0.046455610543489456, -0.0008568777120672166, -0.07026687264442444, 0.20280684530735016, -0.07856161892414093, -0.13774999976158142, -0.13864324986934662, -0.06742197275161743, -0.030877063050866127, 0.03159910440444946, 0.06506045162677765, -0.07244361191987991, -0.05728709325194359, -0.09222450852394104, 0.00902970228344202, 0.03911364823579788, 0.03980075567960739, 0.04425188899040222, 0.0047759306617081165, 0.09709972143173218, -0.10448659956455231, -0.013072875328361988, -0.042155224829912186, -0.04698474705219269, 0.03504662960767746, 0.0454564243555069, 0.11992108821868896, 0.12222544848918915, -0.023811232298612595, -0.0012906916672363877, -0.02173622138798237, 0.23267151415348053, -0.07001236826181412, -0.012684973888099194, 0.1462913453578949, -0.019340990111231804, 0.047758691012859344, 0.1352580189704895, 0.06192997097969055, -0.08839648962020874, 0.021141884848475456, 0.04336819425225258, -0.04039330035448074, -0.20009103417396545, -0.03226656839251518, -0.0460352785885334, -0.016445709392428398, 0.08898679912090302, 0.03216462954878807, 0.04732285067439079, 0.0690927729010582, 0.032611630856990814, 0.0784398764371872, -0.02509250119328499, 0.0733112171292305, 0.08620743453502655, 0.04234786704182625, 0.12356680631637573, -0.04640256240963936, -0.06855650991201401, 0.027605850249528885, 0.003949993755668402, 0.2225702553987503, 0.026926077902317047, 0.11959870159626007, 0.07160840928554535, 0.15747089684009552, -0.012872417457401752, 0.07232402265071869, -0.0038429962005466223, -0.05905373767018318, -0.012359227053821087, -0.05153617262840271, -0.011118040420114994, 0.051628973335027695, -0.08242250233888626, 0.0637424886226654, -0.08743146061897278, 0.04079499468207359, 0.052801914513111115, 0.24585160613059998, 0.054041456431150436, -0.3272058963775635, -0.0949326902627945, 0.022295409813523293, -0.026127852499485016, -0.017812509089708328, 0.018380075693130493, 0.11637930572032928, -0.0511183887720108, 0.041761014610528946, -0.08015809208154678, 0.07084134221076965, 0.011074020527303219, 0.03921077772974968, 0.062319520860910416, 0.10946571826934814, 0.00292105576954782, 0.06453631073236465, -0.2702682912349701, 0.3042501211166382, 0.006979440338909626, 0.09179475158452988, -0.05807296931743622, 0.018449347466230392, 0.03655846789479256, 0.05367451533675194, 0.09952612221240997, -0.021795794367790222, -0.08417796343564987, -0.170880988240242, -0.04296838492155075, 0.022576069459319115, 0.0904308333992958, -0.016502682119607925, 0.10123662650585175, -0.032794542610645294, -0.006213047541677952, 0.09255949407815933, 0.01811593398451805, -0.09348484128713608, -0.08875851333141327, -0.008739843033254147, 0.035643331706523895, -0.06728213280439377, -0.0893629640340805, -0.10255062580108643, -0.11857577413320541, 0.16063790023326874, -0.027040285989642143, -0.021152131259441376, -0.10304298996925354, 0.07386010140180588, 0.06958591192960739, -0.07966821640729904, 0.06946329772472382, 0.009648576378822327, 0.09647415578365326, 0.015085864812135696, -0.05922962725162506, 0.11839598417282104, -0.09442300349473953, -0.154526025056839, -0.07961245626211166, 0.08773846179246902, 0.01645892672240734, 0.05425301566720009, -0.0043234871700406075, 0.02061440236866474, -0.018712962046265602, -0.06536641716957092, 0.04592231661081314, -0.026355480775237083, 0.06106574088335037, 0.012427039444446564, -0.056044138967990875, -0.0045706224627792835, -0.04059883579611778, -0.03322187066078186, 0.14411486685276031, 0.28817179799079895, -0.09399411827325821, 0.006277607288211584, 0.051125574856996536, -0.05738123506307602, -0.21170952916145325, 0.024848561733961105, 0.041159722954034805, 0.011017571203410625, 0.06437952816486359, -0.13895437121391296, 0.11489839851856232, 0.0923314169049263, -0.030463403090834618, 0.10389696061611176, -0.28551429510116577, -0.13723383843898773, 0.13955378532409668, 0.15368624031543732, 0.12060549110174179, -0.14734448492527008, -0.02350062131881714, -0.03892708197236061, -0.11536544561386108, 0.07307103276252747, -0.12344509363174438, 0.11607029289007187, -0.01007026620209217, 0.059962574392557144, -0.002505263779312372, -0.07262669503688812, 0.13118647038936615, -0.02869131602346897, 0.10643316805362701, -0.06552354991436005, -0.0004410663968883455, 0.07370278239250183, -0.04791422188282013, 0.02600139193236828, -0.0991712212562561, 0.03131675720214844, -0.017686884850263596, -0.027791541069746017, -0.05026497691869736, 0.05417871102690697, -0.029046960175037384, -0.07345374673604965, -0.04086935520172119, 0.03200133889913559, 0.03601182997226715, -0.021536553278565407, 0.14728549122810364, 0.03332389518618584, 0.16114717721939087, 0.13048428297042847, 0.048002153635025024, -0.055268410593271255, -0.05184004083275795, -0.0010516769252717495, -0.03544421121478081, 0.06624557077884674, -0.1459326297044754, 0.03218386322259903, 0.10348012298345566, 0.012948628515005112, 0.1467166244983673, 0.07304918020963669, -0.04171615466475487, 0.021430274471640587, 0.08472135663032532, -0.15879812836647034, -0.10353094339370728, 0.0028491520788520575, -0.04958004876971245, -0.1212979331612587, 0.0602102167904377, 0.1135026142001152, -0.07948225736618042, 0.010385500267148018, -0.01209859549999237, 0.020363738760352135, -0.06519882380962372, 0.18788325786590576, 0.06440446525812149, 0.048623308539390564, -0.06548898667097092, 0.07315053790807724, 0.026502346619963646, -0.06488998234272003, 0.018975114449858665, 0.034770067781209946, -0.06342405080795288, -0.03941885009407997, 0.06852222979068756, 0.17883048951625824, -0.02971683070063591, -0.05320744588971138, -0.16099484264850616, -0.10389203578233719, 0.04852927848696709, 0.18166795372962952, 0.09513649344444275, -0.004125901497900486, -0.030664967373013496, 0.033463217318058014, -0.11224007606506348, 0.11093176156282425, 0.04894590377807617, 0.08148828893899918, -0.15351498126983643, 0.11811602115631104, -0.0011830326402559876, 0.02373293600976467, -0.030502179637551308, 0.053649235516786575, -0.10827667266130447, -0.004347928334027529, -0.1160115972161293, -0.021431298926472664, -0.03207928687334061, -0.008097128011286259, 0.012156684882938862, -0.060911621898412704, -0.07489720731973648, 0.028173714876174927, -0.10178538411855698, -0.027922198176383972, 0.04797191917896271, 0.041431039571762085, -0.11995083093643188, -0.04592307284474373, 0.034712426364421844, -0.06242989003658295, 0.053506527096033096, 0.029521776363253593, 0.025868091732263565, 0.04984050989151001, -0.17601452767848969, 0.023769093677401543, 0.0716804638504982, 0.007387302350252867, 0.04685293883085251, -0.1008644849061966, -0.020724069327116013, -0.012895624153316021, 0.047840818762779236, 0.003669239114969969, 0.07286819815635681, -0.12936295568943024, 0.00952344760298729, -0.03472413495182991, -0.07936093956232071, -0.05444050207734108, 0.006143776699900627, 0.0887000560760498, -0.002200116403400898, 0.2057543843984604, -0.1022811159491539, 0.02540190890431404, -0.19899003207683563, 0.01334304641932249, -0.001862578559666872, -0.09616123139858246, -0.10854770988225937, -0.04489592835307121, 0.04451381042599678, -0.06693755835294724, 0.14294663071632385, 0.009998361580073833, 0.01802474819123745, 0.035979460924863815, -0.050032585859298706, 0.047873783856630325, 0.018244987353682518, 0.21934983134269714, 0.017272699624300003, -0.039944469928741455, 0.015795590355992317, 0.03989081829786301, 0.11535444110631943, 0.06471756100654602, 0.17661632597446442, 0.16407550871372223, -0.021349288523197174, 0.11880143731832504, 0.05010131746530533, -0.056329671293497086, -0.14418154954910278, 0.045236147940158844, -0.031399402767419815, 0.09455376863479614, -0.008628502488136292, 0.16963548958301544, 0.12054717540740967, -0.16204380989074707, 0.021188851445913315, -0.04837368428707123, -0.07750510424375534, -0.11558177322149277, -0.06872965395450592, -0.09366319328546524, -0.1417234092950821, 0.015590552240610123, -0.10287365317344666, 0.010920492932200432, 0.08882498741149902, -0.0007117007626220584, -0.014573256485164165, 0.19831515848636627, -0.005989322438836098, 0.04572188854217529, 0.027625085785984993, 0.00674763647839427, -0.037674445658922195, -0.0745306983590126, -0.08148185163736343, -0.010184564627707005, -0.015523560345172882, 0.01625022292137146, -0.061255037784576416, -0.03890227898955345, 0.04988255724310875, -0.002228709403425455, -0.10258571058511734, 0.005259836092591286, 0.022684933617711067, 0.04893791675567627, 0.029489658772945404, 0.004433051683008671, 0.020296936854720116, -0.0062582241371273994, 0.1986991912126541, -0.0833081379532814, -0.08206197619438171, -0.11519229412078857, 0.2041400820016861, 0.026919884607195854, 0.009269779548048973, 0.004267125856131315, -0.08565900474786758, 0.009397432208061218, 0.22973716259002686, 0.20283028483390808, -0.058242157101631165, 0.00677209347486496, 0.003731408854946494, -0.013230190612375736, -0.05760838836431503, 0.08913441747426987, 0.1177084892988205, 0.04006358981132507, -0.07701276242733002, -0.059875406324863434, -0.04636360704898834, -0.0067711276933550835, -0.06186767667531967, 0.03485343977808952, 0.03858282417058945, 0.012620651163160801, -0.03567223250865936, 0.04911774396896362, -0.025623101741075516, -0.11641976237297058, 0.07681098580360413, -0.19262562692165375, -0.14509257674217224, -0.0027081721927970648, 0.12544994056224823, -0.006929036229848862, 0.06346701830625534, -0.03741033375263214, -0.0024700015783309937, 0.07082809507846832, -0.015459422953426838, -0.06813342869281769, -0.08193542063236237, 0.06865989416837692, -0.08846724033355713, 0.2547636926174164, -0.03999466449022293, 0.05667112022638321, 0.1322416216135025, 0.04001326113939285, -0.07330122590065002, 0.086318239569664, 0.049463216215372086, -0.07822772860527039, 0.024519283324480057, 0.10498206317424774, -0.0473339706659317, 0.09853384643793106, 0.049652449786663055, -0.1383780539035797, 0.012163624167442322, -0.054686252027750015, -0.07231393456459045, -0.04196549952030182, -0.020341990515589714, -0.07597159594297409, 0.12930043041706085, 0.18782567977905273, -0.029838483780622482, 0.0013323891907930374, -0.05497472733259201, 0.042751193046569824, 0.08824086934328079, 0.032667845487594604, -0.048462122678756714, -0.23313404619693756, 0.036268994212150574, 0.046638984233140945, -0.018515633419156075, -0.26820093393325806, -0.10490873456001282, -0.009082771837711334, -0.06208930164575577, -0.0845717266201973, 0.0647583082318306, 0.13409389555454254, 0.05809026584029198, -0.05864798277616501, -0.09390919655561447, -0.07490097731351852, 0.1440887302160263, -0.14734813570976257, -0.09988150000572205 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text-generation
unalignment/weeeeee.1
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T16:07:51+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 60, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #conversational #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.04654794931411743, 0.16618601977825165, -0.005445904564112425, 0.01853804849088192, 0.0981811136007309, 0.011998992413282394, 0.06433123350143433, 0.11398410052061081, -0.0230073444545269, 0.11406639218330383, 0.03047988750040531, 0.10172267258167267, 0.11317981779575348, 0.14841650426387787, -0.002152352826669812, -0.22403094172477722, 0.050844956189394, -0.12105348706245422, -0.033293843269348145, 0.11749980598688126, 0.1483822613954544, -0.09928343445062637, 0.07274559140205383, -0.029687678441405296, -0.012143402360379696, -0.030057786032557487, -0.05890674889087677, -0.046214159578084946, 0.04651786759495735, 0.06640566885471344, 0.06770290434360504, 0.0071083661168813705, 0.09012923389673233, -0.2696533799171448, 0.018959321081638336, 0.07145345956087112, -0.002759667346253991, 0.06957992166280746, 0.06404146552085876, -0.07107418030500412, 0.10337356477975845, -0.05106033384799957, 0.14650006592273712, 0.08365883678197861, -0.09081148356199265, -0.1895141303539276, -0.08866965025663376, 0.09882009029388428, 0.17572562396526337, 0.04925641790032387, -0.02320658043026924, 0.09761467576026917, -0.08769196271896362, 0.015438909642398357, 0.04981724172830582, -0.07620415836572647, -0.05378096550703049, 0.05986575037240982, 0.07907199114561081, 0.06627275794744492, -0.12434766441583633, -0.02885502204298973, 0.005009706597775221, 0.010980482213199139, 0.0769270583987236, 0.01728810742497444, 0.146672785282135, 0.0338633768260479, -0.12615777552127838, -0.04880760237574577, 0.09869225323200226, 0.03395522013306618, -0.04422314465045929, -0.24749068915843964, -0.03152675926685333, -0.030810698866844177, -0.029386121779680252, -0.03716538846492767, 0.04340358078479767, -0.007673026993870735, 0.08638741075992584, -0.0060646249912679195, -0.07403432577848434, -0.03937075287103653, 0.06169692054390907, 0.0672287791967392, 0.02999979443848133, -0.013745363801717758, 0.010938193649053574, 0.11620724946260452, 0.1095694974064827, -0.12054188549518585, -0.05555335059762001, -0.06393084675073624, -0.08656639605760574, -0.040790557861328125, 0.034162238240242004, 0.03456587344408035, 0.05349370837211609, 0.25305667519569397, 0.015654386952519417, 0.059652652591466904, 0.034477248787879944, 0.007892133668065071, 0.05848940089344978, 0.11044429242610931, -0.06018859148025513, -0.10444226115942001, -0.02648012898862362, 0.08843598514795303, 0.008199662901461124, -0.03287925571203232, -0.05088530853390694, 0.06019928678870201, 0.01946467161178589, 0.11926145106554031, 0.09061790257692337, 0.010536285117268562, -0.07121123373508453, -0.061038948595523834, 0.1891259253025055, -0.16544590890407562, 0.04322727024555206, 0.035097137093544006, -0.03903156518936157, 0.00019933005387429148, 0.013914269395172596, 0.016625655815005302, -0.025983380153775215, 0.09017423540353775, -0.054113563150167465, -0.04145489260554314, -0.11186197400093079, -0.03383193537592888, 0.033762916922569275, 0.008953776210546494, -0.035059962421655655, -0.033713940531015396, -0.08351044356822968, -0.07577689737081528, 0.09320491552352905, -0.07346344739198685, -0.04878907650709152, -0.01804324984550476, -0.07530532777309418, 0.022395428270101547, 0.019394835457205772, 0.07707412540912628, -0.02362251654267311, 0.04399976506829262, -0.05189276114106178, 0.05863580107688904, 0.11207318305969238, 0.03570080175995827, -0.05736649036407471, 0.06062258034944534, -0.23834340274333954, 0.09552820026874542, -0.07409077137708664, 0.05591456592082977, -0.153293639421463, -0.024439791217446327, 0.04788333550095558, 0.008784620091319084, -0.009650949388742447, 0.13416339457035065, -0.21702027320861816, -0.02536402828991413, 0.1717337965965271, -0.10057014971971512, -0.07069246470928192, 0.05619903281331062, -0.04835370555520058, 0.10988964140415192, 0.03825836628675461, -0.025690359994769096, 0.06171267107129097, -0.1267417073249817, 0.003717758459970355, -0.05005312338471413, -0.017048977315425873, 0.1548657864332199, 0.07182947546243668, -0.07217690348625183, 0.07399354875087738, 0.025708531960844994, -0.0246540866792202, -0.04625825211405754, -0.015164627693593502, -0.10536660254001617, 0.014689887873828411, -0.06369215250015259, 0.014470234513282776, -0.020807426422834396, -0.09071163833141327, -0.027962757274508476, -0.17504668235778809, -0.03014434315264225, 0.08651752024888992, -0.008693269453942776, -0.01803150773048401, -0.1178668737411499, 0.009341353550553322, 0.04177580401301384, 0.0061247628182172775, -0.13462838530540466, -0.04812471568584442, 0.02780051715672016, -0.1600649207830429, 0.034652888774871826, -0.05392369255423546, 0.04932025074958801, 0.025790516287088394, -0.028889117762446404, -0.026493212208151817, 0.021633783355355263, 0.005992184858769178, -0.011999987065792084, -0.24343903362751007, -0.028118690475821495, -0.024888472631573677, 0.1682123839855194, -0.20917098224163055, 0.03546025976538658, 0.07867541164159775, 0.15366052091121674, 0.011240328662097454, -0.04177491366863251, 0.005974748637527227, -0.06935794651508331, -0.02736494317650795, -0.05875484645366669, -0.0047869328409433365, -0.03310677409172058, -0.04545191675424576, 0.04568447172641754, -0.16510973870754242, -0.032636504620313644, 0.09776268899440765, 0.06289951503276825, -0.13922683894634247, -0.020621931180357933, -0.03630133345723152, -0.049253206700086594, -0.04911839962005615, -0.0605199858546257, 0.10893940925598145, 0.05891856551170349, 0.04574795812368393, -0.05928509309887886, -0.07568105310201645, -0.001827909960411489, -0.013898161239922047, -0.017864689230918884, 0.09759635478258133, 0.0751434788107872, -0.13251115381717682, 0.09224759042263031, 0.09603385627269745, 0.07919023185968399, 0.09113933145999908, -0.02355697751045227, -0.08261934667825699, -0.045987509191036224, 0.031442027539014816, 0.020124373957514763, 0.13039541244506836, -0.024294709786772728, 0.04352088272571564, 0.042134687304496765, -0.019369594752788544, 0.014752166345715523, -0.08687400817871094, 0.033972494304180145, 0.028472330421209335, -0.016721390187740326, 0.050190530717372894, -0.03876714035868645, 0.02440318465232849, 0.08830609917640686, 0.045322712510824203, 0.03507532551884651, 0.015493292361497879, -0.05206458270549774, -0.1083620935678482, 0.16405931115150452, -0.12714070081710815, -0.22483378648757935, -0.13936103880405426, 0.0037376401014626026, 0.035628627985715866, -0.015835661441087723, 0.002417160663753748, -0.059374887496232986, -0.12220635265111923, -0.08858037739992142, 0.015140829607844353, 0.04942670464515686, -0.09028962254524231, -0.06437795609235764, 0.058117836713790894, 0.03889724239706993, -0.14560972154140472, 0.017612040042877197, 0.04854894429445267, -0.09789852797985077, -0.006774199660867453, 0.08094939589500427, 0.0698540136218071, 0.1770169734954834, 0.017703235149383545, -0.021850809454917908, 0.032354529947042465, 0.20614571869373322, -0.13538233935832977, 0.11083246022462845, 0.13607586920261383, -0.09041404724121094, 0.08072979003190994, 0.19951270520687103, 0.03932560607790947, -0.10153959691524506, 0.031980328261852264, 0.02283124253153801, -0.0284719280898571, -0.24526868760585785, -0.07212468236684799, -0.004402178805321455, -0.058010730892419815, 0.07660572230815887, 0.09286724030971527, 0.08215958625078201, 0.012304253876209259, -0.09310996532440186, -0.08154371380805969, 0.05942574888467789, 0.10367169976234436, 0.024584239348769188, -0.010839897207915783, 0.08998730033636093, -0.034100502729415894, 0.019626356661319733, 0.0853661298751831, 0.005239574704319239, 0.17840281128883362, 0.05159219726920128, 0.18830420076847076, 0.07925192266702652, 0.07219027727842331, 0.009912233799695969, 0.013080619275569916, 0.018877580761909485, 0.03300119563937187, -0.002769160782918334, -0.08440786600112915, -0.02248465269804001, 0.11566436290740967, 0.06668911874294281, 0.010815348476171494, 0.015172341838479042, -0.04104290530085564, 0.07965951412916183, 0.1831512451171875, -0.007656289264559746, -0.1783534437417984, -0.057547420263290405, 0.07553383708000183, -0.09879875183105469, -0.09854305535554886, -0.013454320840537548, 0.03072015568614006, -0.17046253383159637, 0.023390959948301315, -0.02239842526614666, 0.1106182336807251, -0.14194999635219574, -0.020490378141403198, 0.07218493521213531, 0.07199500501155853, 0.004729843698441982, 0.05758659541606903, -0.16417601704597473, 0.10671813786029816, 0.008950476534664631, 0.06779605895280838, -0.09610627591609955, 0.1008887067437172, -0.004196076653897762, -0.02063460275530815, 0.1393408179283142, 0.002700034761801362, -0.06884108483791351, -0.0763031542301178, -0.08754398673772812, -0.009632662869989872, 0.12754282355308533, -0.1419651061296463, 0.08767123520374298, -0.037212442606687546, -0.0424150750041008, -0.0017086371080949903, -0.10206665843725204, -0.11638247221708298, -0.18888559937477112, 0.06001543253660202, -0.13492922484874725, 0.03152317553758621, -0.10799519717693329, -0.032371897250413895, -0.030304040759801865, 0.19337286055088043, -0.23447458446025848, -0.07199826091527939, -0.1475764364004135, -0.10233612358570099, 0.1443224400281906, -0.0501345656812191, 0.08485390990972519, -0.007241467013955116, 0.16846685111522675, 0.019060896709561348, -0.02531743235886097, 0.0971490666270256, -0.09173708409070969, -0.19302815198898315, -0.07869284600019455, 0.15662524104118347, 0.13260218501091003, 0.031680017709732056, -0.002461588243022561, 0.036563750356435776, -0.015421539545059204, -0.11935004591941833, 0.015969349071383476, 0.1787186712026596, 0.06237189099192619, 0.02331034652888775, -0.027346095070242882, -0.11273157596588135, -0.06900003552436829, -0.028530338779091835, 0.03054865077137947, 0.17762407660484314, -0.07057618349790573, 0.18207968771457672, 0.14163152873516083, -0.05922834202647209, -0.20400173962116241, 0.010538800619542599, 0.03055560030043125, 0.0009220078936778009, 0.02591954916715622, -0.20123432576656342, 0.08688826113939285, 0.004683020059019327, -0.05110127478837967, 0.13194532692432404, -0.17217805981636047, -0.14451217651367188, 0.0765485092997551, 0.038384392857551575, -0.19559739530086517, -0.12913893163204193, -0.09174312651157379, -0.045869920402765274, -0.18591414391994476, 0.09569250047206879, 0.0305706188082695, 0.010893458500504494, 0.03030681423842907, 0.029179483652114868, 0.019487828016281128, -0.0418255440890789, 0.18391458690166473, -0.024792250245809555, 0.026594700291752815, -0.08539514988660812, -0.06927408277988434, 0.03743394836783409, -0.052842434495687485, 0.07349982857704163, -0.023486759513616562, 0.007861839607357979, -0.10348054021596909, -0.042148489505052567, -0.03735732287168503, 0.015448716469109058, -0.09657872468233109, -0.08514349907636642, -0.045032672584056854, 0.09675803780555725, 0.09690850973129272, -0.033646680414676666, -0.028050623834133148, -0.07533035427331924, 0.04412057250738144, 0.19926515221595764, 0.1785389482975006, 0.042153384536504745, -0.08034496754407883, -0.004150947090238333, -0.010121207684278488, 0.04310847446322441, -0.20463712513446808, 0.06283636391162872, 0.05450061708688736, 0.01973269321024418, 0.11436162889003754, -0.019565396010875702, -0.15359151363372803, -0.07263088971376419, 0.06303015351295471, -0.060181066393852234, -0.19620554149150848, 0.00867035984992981, 0.060603946447372437, -0.16371412575244904, -0.04535605385899544, 0.04643881320953369, -0.005620351992547512, -0.038163937628269196, 0.021896906197071075, 0.09194854646921158, 0.0026654244866222143, 0.07427921891212463, 0.05387866869568825, 0.0827430784702301, -0.10537070035934448, 0.08090532571077347, 0.08839722722768784, -0.08452684432268143, 0.023530138656497, 0.10478579998016357, -0.059433579444885254, -0.03440561518073082, 0.020135708153247833, 0.08153781294822693, 0.01775863952934742, -0.040019966661930084, 0.013229827396571636, -0.10452935844659805, 0.05954122915863991, 0.08839859813451767, 0.032507482916116714, 0.016702456399798393, 0.03425082191824913, 0.04607953503727913, -0.07238735258579254, 0.12142276018857956, 0.031868141144514084, 0.017129309475421906, -0.036505792289972305, -0.040896978229284286, 0.019542274996638298, -0.03214648738503456, -0.005015232600271702, -0.03023446537554264, -0.07695909589529037, -0.014793801121413708, -0.1626158058643341, -0.011131818406283855, -0.05648450180888176, 0.010329355485737324, 0.03204665705561638, -0.032609567046165466, 0.008124498650431633, 0.009250079281628132, -0.07695289701223373, -0.0663459524512291, -0.020460480824112892, 0.09540658444166183, -0.16213038563728333, 0.022481130436062813, 0.08244425803422928, -0.12187694013118744, 0.09281346201896667, 0.016204802319407463, -0.006236857734620571, 0.025038830935955048, -0.1475188434123993, 0.034843120723962784, -0.03386561945080757, 0.010836300440132618, 0.04373383894562721, -0.21569781005382538, -0.00004886732858722098, -0.033673107624053955, -0.06639216095209122, -0.009451326914131641, -0.03672455996274948, -0.11508306115865707, 0.1058407872915268, 0.007236586883664131, -0.08753558248281479, -0.03186136856675148, 0.029325377196073532, 0.0838974118232727, -0.021959776058793068, 0.15145497024059296, -0.008370938710868359, 0.07429654151201248, -0.16209737956523895, -0.018623165786266327, -0.006028574425727129, 0.022658247500658035, -0.01664556935429573, -0.01111356820911169, 0.044031109660863876, -0.022746501490473747, 0.17925859987735748, -0.030318550765514374, 0.02272745408117771, 0.06815794110298157, 0.019072026014328003, -0.030184008181095123, 0.10406795144081116, 0.04094860330224037, 0.02014910988509655, 0.018591465428471565, 0.003289656015112996, -0.04647882282733917, -0.03173251822590828, -0.19407226145267487, 0.07288651913404465, 0.15608493983745575, 0.09729263186454773, -0.016707008704543114, 0.07954329252243042, -0.10199416428804398, -0.1109243705868721, 0.12477338314056396, -0.04797708988189697, -0.002418199321255088, -0.07150927931070328, 0.13247236609458923, 0.1437523066997528, -0.1859612911939621, 0.07269313186407089, -0.0699717253446579, -0.04708027467131615, -0.10980689525604248, -0.19441905617713928, -0.05561789125204086, -0.049456022679805756, -0.016053348779678345, -0.04698808491230011, 0.07504211366176605, 0.054538097232580185, 0.006766852922737598, -0.0023397188633680344, 0.06506035476922989, -0.031050674617290497, -0.0037882844917476177, 0.032597362995147705, 0.06591679900884628, 0.012734474614262581, -0.030802709981799126, 0.016619903966784477, -0.013545602560043335, 0.045626189559698105, 0.06578011065721512, 0.04976864159107208, -0.02938537672162056, 0.014603170566260815, -0.038539156317710876, -0.10249634087085724, 0.043612558394670486, -0.024421939626336098, -0.0789753645658493, 0.15477414429187775, 0.023680059239268303, 0.007779473438858986, -0.020137663930654526, 0.23901568353176117, -0.0738423764705658, -0.0964353010058403, -0.14737580716609955, 0.10557299107313156, -0.038081806153059006, 0.05800395458936691, 0.04625935107469559, -0.10226529091596603, 0.018044332042336464, 0.1338089406490326, 0.16182038187980652, -0.039008259773254395, 0.020095856860280037, 0.031135575845837593, 0.00566398398950696, -0.03622615709900856, 0.04847532883286476, 0.06906453520059586, 0.16569648683071136, -0.04632584750652313, 0.09100406616926193, 0.0019041687482967973, -0.09579581767320633, -0.038361791521310806, 0.11069868505001068, -0.016052277758717537, 0.019335128366947174, -0.05818064883351326, 0.11742528527975082, -0.06386786699295044, -0.23783175647258759, 0.06453443318605423, -0.0684293657541275, -0.13765870034694672, -0.02378307841718197, 0.08207765966653824, -0.012955902144312859, 0.027587108314037323, 0.0730307325720787, -0.07240920513868332, 0.201939657330513, 0.03798431158065796, -0.05499868467450142, -0.055047210305929184, 0.0805421993136406, -0.10008571296930313, 0.2739645540714264, 0.01557221356779337, 0.04601577669382095, 0.10384146869182587, -0.009341772645711899, -0.13838784396648407, 0.019836371764540672, 0.09581108391284943, -0.10502193123102188, 0.04196618124842644, 0.19815568625926971, -0.0014755994779989123, 0.12389086186885834, 0.07657600939273834, -0.07551808655261993, 0.0478031262755394, -0.08054235577583313, -0.06760486960411072, -0.09260394424200058, 0.09703279286623001, -0.07772123068571091, 0.14251399040222168, 0.13876807689666748, -0.05074559152126312, 0.012724342755973339, -0.031311117112636566, 0.044293127954006195, -0.00010600237874314189, 0.10321761667728424, 0.004272161517292261, -0.1832672357559204, 0.024692710489034653, 0.005650998093187809, 0.10749758034944534, -0.16033467650413513, -0.09566054493188858, 0.042343202978372574, 0.003505636239424348, -0.0672195628285408, 0.1290110945701599, 0.05665452033281326, 0.04342988133430481, -0.03997718170285225, -0.03521440550684929, -0.0060732318088412285, 0.13561366498470306, -0.10713256150484085, 0.0009933578548952937 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
automatic-speech-recognition
kabir5297/whisper_bn_medium
[ "transformers", "safetensors", "whisper", "automatic-speech-recognition", "arxiv:1910.09700", "endpoints_compatible", "region:us" ]
2024-02-10T16:09:06+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #whisper #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #whisper #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 45, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #whisper #automatic-speech-recognition #arxiv-1910.09700 #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.05983767658472061, 0.15663617849349976, -0.00414510490372777, 0.012621625326573849, 0.10675175487995148, 0.00396517850458622, 0.07058298587799072, 0.10818448662757874, -0.014333043247461319, 0.1301925629377365, 0.031459614634513855, 0.10620059072971344, 0.11486424505710602, 0.17755427956581116, -0.00021593451674561948, -0.21627318859100342, 0.06542544066905975, -0.11467250436544418, 0.023902224376797676, 0.1205042228102684, 0.14280648529529572, -0.10782013833522797, 0.0710505023598671, -0.02651231922209263, -0.014152529649436474, -0.030523719266057014, -0.05870387330651283, -0.06662651896476746, 0.06516408175230026, 0.0716853216290474, 0.05976768955588341, 0.02008269540965557, 0.07725182175636292, -0.2948664724826813, 0.018899710848927498, 0.0727730244398117, 0.011833904311060905, 0.06048334762454033, 0.07948420196771622, -0.06289119273424149, 0.12036014348268509, -0.044804252684116364, 0.1532549113035202, 0.07767832279205322, -0.09226784855127335, -0.19217613339424133, -0.0771055743098259, 0.06758320331573486, 0.1468338817358017, 0.056199874728918076, -0.03856382891535759, 0.15159031748771667, -0.09224481880664825, 0.0102085517719388, 0.06493527442216873, -0.07805083692073822, -0.04958232864737511, 0.027303149923682213, 0.08463363349437714, 0.08637925982475281, -0.1273571401834488, -0.012682586908340454, 0.03438213840126991, 0.02163512259721756, 0.09837246686220169, 0.025364719331264496, 0.11626957356929779, 0.027283066883683205, -0.13964000344276428, -0.055175989866256714, 0.12345059961080551, 0.033505070954561234, -0.05288216099143028, -0.23939087986946106, -0.010561608709394932, -0.009556320495903492, -0.03001241944730282, -0.04216838628053665, 0.03810601681470871, -0.029798293486237526, 0.07650589942932129, 0.01746492274105549, -0.07078345119953156, -0.04342244938015938, 0.06982958316802979, 0.07824850082397461, 0.022348513826727867, -0.02065650187432766, 0.028734240680933, 0.10911912471055984, 0.08262593299150467, -0.12154309451580048, -0.06694398820400238, -0.06854734569787979, -0.09466245025396347, -0.0454239584505558, 0.03469004109501839, 0.06703099608421326, 0.057105712592601776, 0.19864854216575623, 0.011600262485444546, 0.05358051881194115, 0.022981496527791023, 0.01298176683485508, 0.07163717597723007, 0.07945776730775833, -0.051690056920051575, -0.1315721571445465, -0.04847193509340286, 0.11824512481689453, 0.008524151518940926, -0.033710937947034836, -0.02968421019613743, 0.0653507187962532, 0.05568600073456764, 0.11161840707063675, 0.07554161548614502, 0.01568971388041973, -0.07114148139953613, -0.043046265840530396, 0.19346864521503448, -0.15610936284065247, 0.021089470013976097, 0.019353056326508522, -0.05417651683092117, -0.022803083062171936, 0.007743596564978361, 0.017318524420261383, -0.02697303518652916, 0.1045108512043953, -0.07085666805505753, -0.032245416194200516, -0.1046156957745552, -0.055557940155267715, 0.03224421665072441, 0.009115081280469894, -0.030819423496723175, -0.042374368757009506, -0.09924564510583878, -0.0756484866142273, 0.06214139610528946, -0.07012778520584106, -0.06952599436044693, -0.028100011870265007, -0.04856603220105171, 0.012879165820777416, 0.0010717154946178198, 0.12350035458803177, -0.03162076696753502, 0.043779097497463226, -0.04884343594312668, 0.06864890456199646, 0.13179735839366913, 0.032575443387031555, -0.07970008254051208, 0.058469612151384354, -0.22937731444835663, 0.11186469346284866, -0.09973006695508957, 0.03430512547492981, -0.15810096263885498, -0.02635045349597931, 0.024752190336585045, 0.033622484654188156, -0.017231743782758713, 0.13669319450855255, -0.2039388120174408, -0.036121536046266556, 0.1721590757369995, -0.1349588930606842, -0.08518610149621964, 0.06643460690975189, -0.055845119059085846, 0.11782421916723251, 0.049206800758838654, -0.014434589073061943, 0.04594586789608002, -0.13173595070838928, -0.025916490703821182, -0.053098164498806, -0.007177549879997969, 0.15609249472618103, 0.06614800542593002, -0.06571528315544128, 0.03145577386021614, 0.02247771993279457, -0.018577884882688522, -0.045781973749399185, -0.03384651243686676, -0.09418359398841858, 0.007437155116349459, -0.07286001741886139, 0.00992972869426012, -0.017532840371131897, -0.08721724897623062, -0.039823103696107864, -0.16453123092651367, -0.00716154370456934, 0.09300678223371506, 0.010935397818684578, -0.02714768424630165, -0.09726624190807343, 0.006592306774109602, 0.01717078872025013, -0.01454078033566475, -0.15828220546245575, -0.0459267795085907, 0.03719138726592064, -0.1820053607225418, 0.03403490409255028, -0.05244239792227745, 0.035954125225543976, 0.03684226796030998, -0.03816571831703186, -0.013848266564309597, 0.020031210035085678, 0.018333489075303078, -0.017020072788000107, -0.2371053695678711, -0.014824622310698032, -0.04800339788198471, 0.16693253815174103, -0.23147691786289215, 0.03312116861343384, 0.07037223875522614, 0.12888941168785095, 0.003875810420140624, -0.0490296445786953, 0.030063113197684288, -0.05199332535266876, -0.044617995619773865, -0.05644122138619423, -0.006168664898723364, -0.030205117538571358, -0.04949198290705681, 0.050275903195142746, -0.19857677817344666, -0.041567981243133545, 0.11094366759061813, 0.06673718988895416, -0.1588216871023178, -0.0695650652050972, -0.03473977744579315, -0.06271405518054962, -0.09103205800056458, -0.05391426756978035, 0.10852089524269104, 0.04763965308666229, 0.048611950129270554, -0.07248158007860184, -0.04900932312011719, 0.007940629497170448, -0.00704985111951828, -0.03555170074105263, 0.08515505492687225, 0.08571629226207733, -0.11543579399585724, 0.09118600934743881, 0.06718818843364716, 0.06912244111299515, 0.0983632430434227, -0.0017782750073820353, -0.09694159775972366, -0.014548503793776035, 0.018360106274485588, 0.01051856018602848, 0.12805555760860443, -0.07398705929517746, 0.03667636960744858, 0.05262641981244087, -0.035613641142845154, 0.01095122192054987, -0.101106658577919, 0.029197964817285538, 0.0282101072371006, -0.003792217466980219, 0.028733761981129646, -0.04522410035133362, 0.020432880148291588, 0.1023864597082138, 0.03395526856184006, 0.027725959196686745, 0.010809014551341534, -0.04075441509485245, -0.11779133975505829, 0.1720944494009018, -0.09817105531692505, -0.25773105025291443, -0.12466797232627869, -0.001978461164981127, 0.045932475477457047, -0.018764600157737732, 0.01608397625386715, -0.053159136325120926, -0.11253257840871811, -0.10541603714227676, 0.019763922318816185, 0.058765511959791183, -0.08840499073266983, -0.052470505237579346, 0.04951007664203644, 0.036848895251750946, -0.12439411878585815, 0.021039357408881187, 0.04023430123925209, -0.059992119669914246, 0.0014880987582728267, 0.07059671729803085, 0.08472984284162521, 0.18226684629917145, 0.022740190848708153, -0.01784367859363556, 0.017296429723501205, 0.23125670850276947, -0.1456713229417801, 0.09739834815263748, 0.1370985060930252, -0.06344101577997208, 0.08623462915420532, 0.21197044849395752, 0.036558255553245544, -0.08882707357406616, 0.037767693400382996, 0.03336544707417488, -0.036437466740608215, -0.2318716198205948, -0.08410470932722092, 0.001480261329561472, -0.08248372375965118, 0.0952354297041893, 0.09051923453807831, 0.11156398802995682, 0.04929385334253311, -0.10106591880321503, -0.07701091468334198, 0.04251527413725853, 0.11516540497541428, -0.006902680266648531, 0.004321529995650053, 0.09879171848297119, -0.029613742604851723, 0.010339556261897087, 0.09523830562829971, 0.0004232692008372396, 0.18618540465831757, 0.04265686497092247, 0.12916190922260284, 0.08458086103200912, 0.05236417427659035, 0.02661769837141037, 0.01322705764323473, 0.031609587371349335, 0.02576516941189766, -0.02334577962756157, -0.09271565079689026, -0.012906024232506752, 0.1415313482284546, 0.04929639771580696, 0.030407944694161415, 0.020662572234869003, -0.03531459718942642, 0.07301895320415497, 0.16116659343242645, 0.011933310888707638, -0.21851851046085358, -0.05515235662460327, 0.07743874937295914, -0.08626089245080948, -0.11299191415309906, -0.0025294655933976173, 0.021754881367087364, -0.17833879590034485, 0.05397404730319977, -0.016486117616295815, 0.10160378366708755, -0.11242987960577011, -0.02206907607614994, 0.04055493697524071, 0.07460751384496689, -0.03305850550532341, 0.07621917128562927, -0.20276865363121033, 0.1373196691274643, 0.008098544552922249, 0.06249339506030083, -0.11230216175317764, 0.08414414525032043, 0.019059745594859123, -0.0036223498173058033, 0.1621086448431015, -0.009664713405072689, -0.09406581521034241, -0.060111574828624725, -0.07602227479219437, -0.012445085681974888, 0.09843466430902481, -0.0939253643155098, 0.08608877658843994, -0.01022840291261673, -0.03214890882372856, -0.007143673487007618, -0.11786875873804092, -0.1394684612751007, -0.183831125497818, 0.05997816100716591, -0.10696699470281601, 0.03344186022877693, -0.10895431786775589, -0.060553617775440216, -0.03646453842520714, 0.19020794332027435, -0.18181639909744263, -0.08386372029781342, -0.14476649463176727, -0.07653295993804932, 0.1361350119113922, -0.04076695069670677, 0.07850751280784607, -0.00008746175444684923, 0.20719517767429352, 0.001825421117246151, -0.00039511307841166854, 0.08349475264549255, -0.09573810547590256, -0.20032998919487, -0.0880952924489975, 0.13964824378490448, 0.12494690716266632, 0.04542626440525055, -0.006928097922354937, 0.027518225833773613, -0.011671899817883968, -0.11464269459247589, 0.02507087029516697, 0.1405206173658371, 0.06840235739946365, 0.04314489662647247, -0.016979211941361427, -0.15606153011322021, -0.10666806995868683, -0.05322869494557381, 0.021586019545793533, 0.17797614634037018, -0.07007403671741486, 0.1621050238609314, 0.16129834949970245, -0.05420130863785744, -0.2030099630355835, 0.02282964438199997, 0.04042449966073036, -0.013990761712193489, 0.03615177795290947, -0.19683793187141418, 0.07753707468509674, 0.016794858500361443, -0.060990821570158005, 0.13549083471298218, -0.1619698405265808, -0.1508903205394745, 0.09218499809503555, 0.06408262252807617, -0.2138945758342743, -0.13302136957645416, -0.10209991782903671, -0.05448025092482567, -0.10983701795339584, 0.08582660555839539, 0.01998555287718773, 0.0000906725981622003, 0.04219266399741173, 0.03161109238862991, 0.021054213866591454, -0.0520465187728405, 0.20073460042476654, 0.0012120193568989635, 0.03459459915757179, -0.08232162147760391, -0.08637090027332306, 0.026973288506269455, -0.05251563340425491, 0.0672052875161171, -0.016655180603265762, 0.0002542635484132916, -0.09922616183757782, -0.06439188867807388, -0.06020424887537956, 0.03343502804636955, -0.08179902285337448, -0.09706422686576843, -0.058388181030750275, 0.10227678716182709, 0.08968468755483627, -0.03377925977110863, -0.06091363728046417, -0.10292473435401917, 0.06651771068572998, 0.22872710227966309, 0.1885143369436264, 0.06312023848295212, -0.07107747346162796, 0.0009368667961098254, -0.023646708577871323, 0.050360288470983505, -0.1945972442626953, 0.046965986490249634, 0.042262639850378036, 0.028454279527068138, 0.12927067279815674, -0.024874795228242874, -0.16607771813869476, -0.04733136296272278, 0.06063033267855644, -0.059542834758758545, -0.18076083064079285, -0.000619421829469502, 0.09315520524978638, -0.15953904390335083, -0.06748805940151215, 0.023891208693385124, -0.020897341892123222, -0.027535755187273026, 0.004573860205709934, 0.0820559412240982, 0.02817925252020359, 0.11291294544935226, 0.06535529345273972, 0.10744494199752808, -0.10965088754892349, 0.08151662349700928, 0.09152320772409439, -0.10730767250061035, 0.02777967043220997, 0.07435369491577148, -0.05882004648447037, -0.03269755467772484, 0.0057791233994066715, 0.07514561712741852, 0.02294853888452053, -0.07087770849466324, -0.0009696646011434495, -0.1182747483253479, 0.06833867728710175, 0.13341592252254486, 0.033248964697122574, -0.0019442925695329905, 0.044254120439291, 0.02532937377691269, -0.08849740773439407, 0.11402047425508499, 0.03831348940730095, 0.031180279329419136, -0.04628003388643265, -0.005872894544154406, 0.04073992744088173, -0.011434492655098438, -0.01770744100213051, -0.03857431188225746, -0.061015255749225616, -0.009887747466564178, -0.1567201316356659, 0.02684243768453598, -0.0771009624004364, 0.00816130917519331, 0.022786233574151993, -0.03996667265892029, -0.005420312751084566, 0.006734060123562813, -0.08264576643705368, -0.03730582818388939, -0.0037628922145813704, 0.1070059984922409, -0.15296638011932373, 0.00852613802999258, 0.09225248545408249, -0.12423861026763916, 0.07808402180671692, -0.0011087276507169008, -0.013306759297847748, 0.02074836567044258, -0.1374569684267044, 0.051461800932884216, -0.006391053553670645, 0.011301612481474876, 0.028202330693602562, -0.19194763898849487, 0.0008063786081038415, -0.04062483087182045, -0.05044460669159889, -0.012731820344924927, -0.05135709419846535, -0.11374296247959137, 0.10732509195804596, 0.023315785452723503, -0.08887150883674622, -0.01889934204518795, 0.045546844601631165, 0.10550197213888168, -0.05122669041156769, 0.13676951825618744, -0.01927841641008854, 0.0586048886179924, -0.1769271343946457, -0.014012092724442482, -0.018402719870209694, 0.013554446399211884, -0.017449822276830673, -0.00605781190097332, 0.0551704466342926, -0.012471658177673817, 0.23972837626934052, -0.027916517108678818, 0.03500373288989067, 0.06697984784841537, 0.016924316063523293, -0.018179070204496384, 0.08486920595169067, 0.05455834046006203, 0.026243781670928, 0.01494054775685072, 0.017568159848451614, -0.051871586591005325, -0.021555433049798012, -0.1424977034330368, 0.07956096529960632, 0.16729016602039337, 0.09009124338626862, -0.008234765380620956, 0.06473081558942795, -0.11607895791530609, -0.07983584702014923, 0.10896016657352448, -0.03711748123168945, -0.0032444922253489494, -0.05700715631246567, 0.1502007693052292, 0.1525147259235382, -0.16814833879470825, 0.06879524886608124, -0.06271831691265106, -0.05224054306745529, -0.11435537785291672, -0.16904489696025848, -0.06866718828678131, -0.035694681107997894, -0.002330650808289647, -0.05624498426914215, 0.07767387479543686, 0.10255347937345505, 0.007528870366513729, 0.0038026864640414715, 0.08233556896448135, -0.037537459284067154, -0.006316144950687885, 0.04542352631688118, 0.049430496990680695, 0.015805410221219063, -0.059124622493982315, 0.010986202396452427, 0.004953318741172552, 0.04692067950963974, 0.05509426072239876, 0.034005217254161835, -0.028324270620942116, 0.012686561793088913, -0.018243486061692238, -0.10028578341007233, 0.035927701741456985, -0.033664118498563766, -0.05780354142189026, 0.13973994553089142, 0.0218597874045372, 0.007779987063258886, -0.02196359448134899, 0.22996114194393158, -0.07252145558595657, -0.08971016108989716, -0.1408918797969818, 0.13730354607105255, -0.046912964433431625, 0.05402535945177078, 0.04905577376484871, -0.10465127229690552, 0.0241316556930542, 0.14292258024215698, 0.13702698051929474, -0.027644719928503036, 0.010874779894948006, 0.015687033534049988, 0.00620539765805006, -0.031101418659090996, 0.04872303828597069, 0.04169761762022972, 0.13120494782924652, -0.06359384953975677, 0.0914405807852745, -0.010274309664964676, -0.08765450119972229, -0.0231675673276186, 0.1299583613872528, 0.005232672207057476, 0.02307419292628765, -0.08125553280115128, 0.11583263427019119, -0.0691702738404274, -0.24996554851531982, 0.04865904897451401, -0.05924736708402634, -0.15156961977481842, -0.017320360988378525, 0.02757420763373375, 0.005632835440337658, 0.02303774654865265, 0.06296881288290024, -0.06651590019464493, 0.1557060331106186, 0.035915885120630264, -0.07977382838726044, -0.06385304778814316, 0.08052598685026169, -0.08511006832122803, 0.29178112745285034, 0.010383724234998226, 0.05882499739527702, 0.0948280319571495, -0.028215935453772545, -0.131154403090477, 0.05278646945953369, 0.0955355316400528, -0.07669185847043991, 0.070269875228405, 0.19858962297439575, 0.0003398389380890876, 0.11546503752470016, 0.07913552224636078, -0.09058261662721634, 0.05968843400478363, -0.07367776334285736, -0.09094593673944473, -0.0922231450676918, 0.08578167855739594, -0.06759190559387207, 0.15170368552207947, 0.12874077260494232, -0.043129127472639084, -0.001158626051619649, -0.030579449608922005, 0.051351167261600494, -0.0008969766786321998, 0.12188339978456497, 0.015837527811527252, -0.19386562705039978, 0.031386423856019974, -0.015537483617663383, 0.099497489631176, -0.23898114264011383, -0.07769263535737991, 0.03750690072774887, -0.014490727335214615, -0.048680152744054794, 0.11743341386318207, 0.05373985692858696, 0.045937854796648026, -0.05465031415224075, -0.060885775834321976, 0.006574091035872698, 0.1611197590827942, -0.11137263476848602, 0.004140520468354225 ]
null
null
null
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # distilbert-base-uncased-lora-text-classification This model is a fine-tuned version of [distilbert-base-uncased](https://huggingface.co/distilbert-base-uncased) on the None dataset. It achieves the following results on the evaluation set: - Loss: 1.3834 - Precision: 0.8310 - Recall: 0.8708 - F1 and accuracy: {'accuracy': 0.7877237851662404, 'f1': 0.8504504504504504} ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.001 - train_batch_size: 4 - eval_batch_size: 4 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 10 ### Training results | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 and accuracy | |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:----------------------------------------------------------:| | No log | 1.0 | 391 | 0.5803 | 0.7346 | 0.9705 | {'accuracy': 0.7365728900255755, 'f1': 0.836248012718601} | | 0.5606 | 2.0 | 782 | 0.5085 | 0.8259 | 0.8229 | {'accuracy': 0.7570332480818415, 'f1': 0.8243992606284658} | | 0.4687 | 3.0 | 1173 | 0.6925 | 0.8007 | 0.8745 | {'accuracy': 0.7621483375959079, 'f1': 0.8359788359788359} | | 0.3603 | 4.0 | 1564 | 0.8182 | 0.7955 | 0.9188 | {'accuracy': 0.7800511508951407, 'f1': 0.8527397260273973} | | 0.3603 | 5.0 | 1955 | 0.8375 | 0.8413 | 0.8413 | {'accuracy': 0.7800511508951407, 'f1': 0.8413284132841329} | | 0.2736 | 6.0 | 2346 | 1.0186 | 0.8235 | 0.8782 | {'accuracy': 0.7851662404092071, 'f1': 0.8500000000000001} | | 0.1993 | 7.0 | 2737 | 1.1566 | 0.8224 | 0.9225 | {'accuracy': 0.8081841432225064, 'f1': 0.8695652173913043} | | 0.1491 | 8.0 | 3128 | 1.2136 | 0.8502 | 0.8376 | {'accuracy': 0.7851662404092071, 'f1': 0.8438661710037174} | | 0.1224 | 9.0 | 3519 | 1.3815 | 0.8231 | 0.8930 | {'accuracy': 0.7928388746803069, 'f1': 0.8566371681415929} | | 0.1224 | 10.0 | 3910 | 1.3834 | 0.8310 | 0.8708 | {'accuracy': 0.7877237851662404, 'f1': 0.8504504504504504} | ### Framework versions - Transformers 4.35.2 - Pytorch 2.1.0+cu121 - Datasets 2.17.0 - Tokenizers 0.15.1
{"license": "apache-2.0", "tags": ["generated_from_trainer"], "metrics": ["precision", "recall"], "base_model": "distilbert-base-uncased", "model-index": [{"name": "distilbert-base-uncased-lora-text-classification", "results": []}]}
null
nhankins/en_distil_lora_adapter_1.0
[ "tensorboard", "safetensors", "generated_from_trainer", "base_model:distilbert-base-uncased", "license:apache-2.0", "region:us" ]
2024-02-10T16:09:45+00:00
[]
[]
TAGS #tensorboard #safetensors #generated_from_trainer #base_model-distilbert-base-uncased #license-apache-2.0 #region-us
distilbert-base-uncased-lora-text-classification ================================================ This model is a fine-tuned version of distilbert-base-uncased on the None dataset. It achieves the following results on the evaluation set: * Loss: 1.3834 * Precision: 0.8310 * Recall: 0.8708 * F1 and accuracy: {'accuracy': 0.7877237851662404, 'f1': 0.8504504504504504} Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 0.001 * train\_batch\_size: 4 * eval\_batch\_size: 4 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 10 ### Training results ### Framework versions * Transformers 4.35.2 * Pytorch 2.1.0+cu121 * Datasets 2.17.0 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10", "### Training results", "### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ "TAGS\n#tensorboard #safetensors #generated_from_trainer #base_model-distilbert-base-uncased #license-apache-2.0 #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10", "### Training results", "### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ 44, 97, 4, 33 ]
[ "passage: TAGS\n#tensorboard #safetensors #generated_from_trainer #base_model-distilbert-base-uncased #license-apache-2.0 #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 4\n* eval\\_batch\\_size: 4\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 10### Training results### Framework versions\n\n\n* Transformers 4.35.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.17.0\n* Tokenizers 0.15.1" ]
[ -0.10776074975728989, 0.013286866247653961, -0.0009072066168300807, 0.10292543470859528, 0.1671837866306305, 0.015910841524600983, 0.14432811737060547, 0.07169800996780396, -0.1033245325088501, 0.03434503450989723, 0.10547147691249847, 0.1356198787689209, -0.00041341534233652055, 0.08636415004730225, -0.06794607639312744, -0.18573631346225739, 0.006940762046724558, 0.010847656987607479, -0.07067512720823288, 0.1084565818309784, 0.0819510743021965, -0.1482250839471817, 0.062359970062971115, -0.024163924157619476, -0.2240094095468521, 0.034070488065481186, 0.03639496862888336, -0.04664241522550583, 0.1320042759180069, -0.018770040944218636, 0.1575230211019516, 0.004414635710418224, 0.11843379586935043, -0.1664608269929886, 0.013683661818504333, 0.08807466179132462, 0.004451241344213486, 0.07409375160932541, 0.05300283059477806, -0.0032721247989684343, 0.06854736804962158, -0.11246369779109955, 0.06507588177919388, 0.024089861661195755, -0.12630298733711243, -0.25454673171043396, -0.10020113736391068, -0.0019044613000005484, 0.07107430696487427, 0.08783601969480515, -0.00902322307229042, 0.18306183815002441, -0.07895683497190475, 0.0809808298945427, 0.23842281103134155, -0.2897263765335083, -0.07235468178987503, 0.06325308233499527, 0.004995733965188265, 0.10359063744544983, -0.10336722433567047, -0.017051314935088158, 0.08697554469108582, 0.039989203214645386, 0.11872565746307373, -0.0253953505307436, -0.16234475374221802, 0.023050833493471146, -0.15416018664836884, 0.027474647387862206, 0.08527977764606476, 0.041305236518383026, -0.04898786544799805, 0.014076385647058487, -0.07982347905635834, -0.10469923913478851, -0.05187740921974182, -0.03556440398097038, 0.07759159058332443, -0.045776885002851486, -0.07574113458395004, -0.005737872328609228, -0.09976216405630112, -0.09119320660829544, -0.050533656030893326, 0.20141872763633728, 0.05305128172039986, 0.04221257567405701, -0.03722304850816727, 0.10384555160999298, -0.06541979312896729, -0.13885870575904846, 0.04070119187235832, 0.030343685299158096, 0.000685641192831099, -0.052194975316524506, -0.06724326312541962, -0.07755634188652039, 0.039889611303806305, 0.140565887093544, -0.11287222057580948, 0.036636367440223694, 0.02240842394530773, 0.041084129363298416, -0.1139121800661087, 0.1371600329875946, -0.043160825967788696, 0.005759380292147398, 0.037347402423620224, 0.07309786975383759, 0.054738447070121765, 0.0009091220563277602, -0.08039036393165588, 0.04318290203809738, 0.08723826706409454, 0.019060250371694565, -0.07131239771842957, 0.0351664237678051, -0.04042354226112366, 0.014625527895987034, -0.00013166219287086278, -0.10735303908586502, 0.04139012470841408, 0.008346169255673885, -0.07553161680698395, -0.032856110483407974, 0.023005478084087372, 0.028185592964291573, 0.018932217732071877, 0.11620213091373444, -0.08295289427042007, 0.05272528901696205, -0.11641733348369598, -0.11223863065242767, 0.0006384730222634971, -0.025775859132409096, 0.018781783059239388, -0.08522629737854004, -0.17215797305107117, -0.01021167729049921, 0.06536132842302322, -0.028724953532218933, 0.0022211414761841297, -0.03773406520485878, -0.09885777533054352, -0.008485173806548119, -0.015379080548882484, 0.13597020506858826, -0.06201976537704468, 0.09089437127113342, 0.0645659789443016, 0.058013781905174255, -0.09662272036075592, 0.022656302899122238, -0.11202622205018997, 0.005841631442308426, -0.24800796806812286, -0.008465999737381935, -0.08178983628749847, 0.07737165689468384, -0.05676591023802757, -0.06675288081169128, 0.012126335874199867, -0.0013353358954191208, 0.1074616089463234, 0.09710410237312317, -0.20027033984661102, -0.054181188344955444, 0.13845223188400269, -0.09803525358438492, -0.124534472823143, 0.10040358453989029, -0.0509214773774147, 0.04818853363394737, 0.07998498529195786, 0.17882923781871796, -0.019456036388874054, -0.12007530778646469, 0.0048804678954184055, -0.0302610881626606, 0.06125456467270851, -0.05798470228910446, 0.04107387736439705, -0.018492968752980232, -0.014312369748950005, 0.02435508742928505, -0.05744905024766922, 0.032609567046165466, -0.11124904453754425, -0.08824419230222702, -0.07714127749204636, -0.11267666518688202, 0.0192737877368927, 0.05953573063015938, 0.06410989910364151, -0.12365350872278214, -0.06626294553279877, 0.10229677706956863, 0.08024650067090988, -0.05848350375890732, 0.027403371408581734, -0.03802352771162987, 0.07155103981494904, -0.06315396726131439, -0.03223726898431778, -0.1729402393102646, -0.05951324477791786, 0.010662139393389225, -0.008625210262835026, 0.027479596436023712, -0.03213559836149216, 0.0803985670208931, 0.09031280875205994, -0.06969557702541351, -0.010902360081672668, -0.05358225107192993, 0.014028947800397873, -0.12810580432415009, -0.24172891676425934, -0.0081186443567276, -0.021048355847597122, 0.08609570562839508, -0.2101007103919983, 0.0348973274230957, -0.05878216028213501, 0.0833989754319191, 0.01658683642745018, -0.009553615003824234, -0.0633246973156929, 0.08538926392793655, -0.004885191563516855, -0.06906567513942719, 0.04060029238462448, -0.011504565365612507, -0.06696385890245438, -0.0891338661313057, -0.10470440238714218, 0.1690089851617813, 0.14966942369937897, -0.11664288491010666, -0.07864658534526825, 0.048693522810935974, -0.05569438263773918, -0.01996980421245098, -0.0706285759806633, 0.053830720484256744, 0.13230887055397034, -0.012343679554760456, 0.11829325556755066, -0.08924629539251328, -0.03019736148416996, 0.00030790732125751674, -0.04702973738312721, 0.07185359299182892, 0.08079186826944351, 0.14181484282016754, -0.05704277381300926, 0.1173090860247612, 0.19219493865966797, -0.13252170383930206, 0.08345472812652588, -0.06926695257425308, -0.06929399818181992, -0.015585428103804588, 0.0017115407390519977, -0.004419997334480286, 0.1763026863336563, -0.05474870651960373, 0.027019023895263672, -0.010043533518910408, 0.012866819277405739, 0.02480131760239601, -0.25005653500556946, -0.04346234351396561, -0.003466873662546277, -0.04195188358426094, -0.021398449316620827, -0.04102545231580734, 0.006357361096888781, 0.10497906804084778, -0.04747966676950455, -0.07744108140468597, 0.02068677358329296, 0.006034460850059986, -0.06950552016496658, 0.21867947280406952, -0.08889976888895035, -0.020531799644231796, -0.062195856124162674, -0.00925856176763773, -0.05387485772371292, -0.005203715525567532, 0.04962547495961189, -0.11158133298158646, -0.04353714361786842, -0.08091434091329575, -0.0021962313912808895, 0.06762607395648956, 0.03837525472044945, 0.026805907487869263, -0.004182002507150173, 0.11132576316595078, -0.1282758116722107, 0.0061349631287157536, -0.07833211869001389, -0.08436174690723419, 0.031129198148846626, 0.10977601259946823, 0.10885851085186005, 0.15138964354991913, -0.02330865152180195, -0.018250077962875366, -0.010055607184767723, 0.2502712309360504, -0.050127558410167694, -0.0334346778690815, 0.10562203824520111, -0.006119457073509693, 0.051463425159454346, 0.1196419820189476, 0.10133781284093857, -0.12179071456193924, 0.010243627242743969, 0.048708926886320114, -0.022350329905748367, -0.23310470581054688, -0.02446550875902176, -0.01910938136279583, -0.07277027517557144, 0.04593532904982567, 0.033917468041181564, 0.006453594192862511, 0.05901132524013519, 0.037824392318725586, 0.06853179633617401, -0.04635021463036537, 0.046201303601264954, 0.04422686994075775, 0.05320153385400772, 0.1106828823685646, -0.053859319537878036, -0.05316624790430069, 0.0305063147097826, -0.0352352000772953, 0.2397363930940628, 0.010504846461117268, 0.06275320798158646, 0.0810859426856041, 0.19888019561767578, -0.028027545660734177, 0.08461993932723999, -0.0002573218662291765, -0.07118545472621918, 0.003677784465253353, -0.07169575989246368, 0.020682236179709435, 0.017586655914783478, -0.14050410687923431, 0.0880737230181694, -0.09226199984550476, 0.0005032197223044932, 0.08365349471569061, 0.223918154835701, 0.028514623641967773, -0.3037088215351105, -0.05156293511390686, 0.0056722271256148815, -0.001373344217427075, 0.0013688887702301145, 0.01118419412523508, 0.1202123761177063, -0.017463993281126022, 0.03791265934705734, -0.058203861117362976, 0.07585317641496658, 0.0419001504778862, 0.04391868785023689, 0.0517541877925396, 0.15626682341098785, -0.011268813163042068, 0.04026207700371742, -0.2882959842681885, 0.2711438834667206, 0.03184324502944946, 0.13550254702568054, -0.03653101623058319, -0.011836283840239048, 0.03771466016769409, 0.059997204691171646, 0.028666943311691284, -0.02172599546611309, -0.06370425224304199, -0.16935954988002777, -0.04577065631747246, 0.06437711417675018, 0.11465051770210266, 0.02251414768397808, 0.09301172196865082, -0.00660290801897645, 0.00816494133323431, 0.10379992425441742, -0.052157845348119736, -0.17178691923618317, -0.037516094744205475, -0.05926873907446861, 0.03499923646450043, -0.0884702131152153, -0.0844932571053505, -0.1030658707022667, -0.1347644031047821, 0.11006185412406921, 0.007308953907340765, -0.018469953909516335, -0.10076653212308884, 0.08476575464010239, 0.0841209888458252, -0.052141223102808, 0.04484931007027626, 0.026690857484936714, 0.03123578242957592, 0.032211508601903915, -0.0665707215666771, 0.11320937424898148, -0.07477326691150665, -0.16500522196292877, -0.05994654819369316, 0.06980053335428238, 0.04711758345365524, 0.037371281534433365, -0.017458293586969376, 0.019877947866916656, -0.011363241821527481, -0.1029902771115303, 0.021586544811725616, -0.024103721603751183, 0.05588429793715477, 0.0238752793520689, -0.05686929076910019, -0.024303283542394638, -0.05611411854624748, -0.02567092329263687, 0.10586445033550262, 0.3159607946872711, -0.06811869144439697, -0.017014222219586372, 0.07098253816366196, -0.05378156155347824, -0.18034197390079498, 0.09635962545871735, 0.03581840544939041, -0.016717830672860146, 0.08179271966218948, -0.12879784405231476, 0.17212818562984467, 0.12054923176765442, -0.027904052287340164, 0.13846451044082642, -0.29091933369636536, -0.14509937167167664, 0.11415629088878632, 0.19187107682228088, 0.1522778570652008, -0.16437450051307678, -0.02670810930430889, -0.03505118191242218, -0.08123420923948288, 0.10132340341806412, -0.2278299480676651, 0.0855850875377655, -0.0013330637011677027, 0.06309229880571365, -0.008642949163913727, -0.06039194390177727, 0.14621402323246002, -0.0010026143863797188, 0.15335923433303833, -0.04101422801613808, 0.00942506268620491, 0.08330554515123367, -0.02939201146364212, 0.012100488878786564, -0.07467890530824661, 0.04487605765461922, 0.018958428874611855, -0.0003758893290068954, -0.06895087659358978, 0.05006789788603783, -0.04415126517415047, -0.0631125420331955, -0.038396481424570084, 0.009171977639198303, 0.008995141834020615, -0.02835332229733467, 0.11959891021251678, 0.049258872866630554, 0.16831055283546448, 0.09512530267238617, 0.026240166276693344, -0.0866500660777092, -0.01615576446056366, 0.02772725373506546, -0.03400951251387596, 0.07071332633495331, -0.16963687539100647, 0.01726691983640194, 0.1082320287823677, 0.023486752063035965, 0.09463851898908615, 0.07905522733926773, -0.05509939789772034, 0.009418448433279991, 0.07202079147100449, -0.16806840896606445, -0.0880303829908371, 0.03782069683074951, -0.04245994985103607, -0.08835507929325104, 0.10541251301765442, 0.08151200413703918, -0.07815353572368622, -0.0012020826106891036, -0.018748989328742027, 0.003005637787282467, -0.07947627454996109, 0.2218736857175827, 0.09277518838644028, 0.03656836599111557, -0.10076059401035309, 0.10971776396036148, 0.03936690092086792, -0.04275985434651375, -0.0059500085189938545, 0.047347117215394974, -0.062312640249729156, -0.011678621172904968, 0.15288057923316956, 0.23133595287799835, -0.0341607965528965, -0.061460357159376144, -0.17149095237255096, -0.10264989733695984, 0.03005433827638626, 0.18359892070293427, 0.10098309069871902, -0.014587575569748878, 0.001004935009405017, 0.03358463570475578, -0.12887556850910187, 0.08343463391065598, 0.031826265156269073, 0.08621089905500412, -0.13972941040992737, 0.12862944602966309, 0.02536475472152233, 0.0018114804988726974, -0.024594802409410477, 0.07080560177564621, -0.12159477174282074, 0.016615893691778183, -0.14901386201381683, -0.04314075782895088, 0.010290384292602539, -0.00662763649597764, 0.017358576878905296, -0.07121779769659042, -0.08295535296201706, 0.04020083323121071, -0.11930908262729645, -0.016839463263750076, 0.0625620186328888, 0.02549903653562069, -0.1588132232427597, -0.02951795607805252, 0.01745225302875042, -0.05555880442261696, 0.024031205102801323, 0.04840434342622757, 0.023638786748051643, 0.09720733761787415, -0.244621142745018, -0.014732897281646729, 0.0722523033618927, -0.001762350439094007, 0.07911508530378342, -0.04315507411956787, -0.04027720168232918, 0.003731303382664919, 0.10443970561027527, 0.008588923141360283, 0.08916698396205902, -0.12514959275722504, -0.005229677073657513, -0.04356436803936958, -0.07321158796548843, -0.034577127546072006, -0.010922924615442753, 0.08413414657115936, -0.0017459566006436944, 0.1957140415906906, -0.0941910669207573, 0.01010021660476923, -0.22430220246315002, -0.004233746789395809, -0.021257178857922554, -0.08863459527492523, -0.13560622930526733, -0.019329287111759186, 0.06740187853574753, -0.060044627636671066, 0.1443650722503662, 0.0034738394897431135, 0.0562247596681118, 0.03301425650715828, -0.01843157596886158, 0.0031336003448814154, 0.041586846113204956, 0.24669817090034485, 0.03193332627415657, -0.008837562054395676, 0.018831131979823112, 0.05767788365483284, 0.11772309243679047, 0.03452807292342186, 0.2190881371498108, 0.180728018283844, -0.044404201209545135, 0.10383239388465881, 0.05004359781742096, -0.07724038511514664, -0.08630583435297012, 0.0728040412068367, -0.04136200621724129, 0.06032751873135567, -0.022283639758825302, 0.19778037071228027, 0.08856140077114105, -0.15187646448612213, 0.02263619191944599, -0.0588926300406456, -0.08297762274742126, -0.11527875065803528, 0.02511155605316162, -0.07880032062530518, -0.1867126226425171, 0.016939349472522736, -0.10100153833627701, 0.009158856235444546, 0.1337585598230362, 0.0011021103709936142, -0.004866723902523518, 0.2119547575712204, 0.05624644458293915, 0.06294143199920654, 0.025914497673511505, -0.0003623918164521456, -0.05397367104887962, -0.061271920800209045, -0.09891465306282043, 0.0007082002121023834, -0.020740462467074394, 0.027084186673164368, -0.06044841185212135, -0.08244873583316803, 0.06489557027816772, 0.004191926680505276, -0.0970083624124527, 0.030736343935132027, 0.02022668719291687, 0.05680035054683685, 0.03483536094427109, 0.01650865562260151, 0.021458227187395096, -0.010838502086699009, 0.22482040524482727, -0.07923631370067596, -0.11427892744541168, -0.08007156848907471, 0.23739713430404663, 0.023823419585824013, -0.0015932457754388452, 0.011138982139527798, -0.10882855206727982, 0.014431757852435112, 0.180855393409729, 0.17740124464035034, -0.11697714030742645, -0.0075906189158558846, -0.04016681760549545, -0.02286985144019127, -0.0983743667602539, 0.11888343840837479, 0.126287043094635, 0.00635329307988286, -0.10206305980682373, -0.042170990258455276, -0.04955526813864708, -0.008839772082865238, -0.07488121837377548, 0.025880437344312668, 0.032449085265398026, 0.009460116736590862, -0.05012940987944603, 0.072842538356781, -0.019152237102389336, -0.13213934004306793, 0.0705651044845581, -0.15992075204849243, -0.16074852645397186, -0.022876614704728127, 0.15635506808757782, -0.01981372758746147, 0.048732955008745193, -0.04774896800518036, 0.0017805523239076138, 0.05618862435221672, -0.04297444224357605, -0.048742447048425674, -0.12208903580904007, 0.0781761035323143, -0.13913196325302124, 0.2427528202533722, -0.022185906767845154, 0.10440786927938461, 0.11486203968524933, 0.03449452668428421, -0.0749020054936409, 0.09630509465932846, 0.03572151064872742, -0.1332705318927765, 0.0060898312367498875, 0.0821838527917862, -0.04786362126469612, 0.0620090551674366, 0.03672077879309654, -0.10733720660209656, 0.01617717556655407, -0.011213628575205803, -0.0813797116279602, -0.0573691800236702, -0.0665663480758667, -0.07499007135629654, 0.10478900372982025, 0.1638081967830658, -0.027784906327724457, 0.06208745762705803, -0.07081511616706848, 0.052178774029016495, 0.06718576699495316, 0.036514464765787125, -0.047597236931324005, -0.25648635625839233, 0.07059800624847412, 0.12126485258340836, -0.03756977617740631, -0.22011439502239227, -0.0767139121890068, 0.022021662443876266, -0.06884575635194778, -0.10017000138759613, 0.059786535799503326, 0.13267862796783447, 0.061204519122838974, -0.0562906414270401, -0.153342604637146, -0.09310054033994675, 0.15074148774147034, -0.1274908035993576, -0.10189483314752579 ]
null
null
transformers
![Reyna aloobun qwen0.5B](https://i.imgur.com/QfbOY6c.jpeg) - Finetuned [Qwen/Qwen1.5-0.5B-Chat](https://huggingface.co/Qwen/Qwen1.5-0.5B-Chat) over 3 epochs, made primarily for RP (Roleplay) and/or storywriting. - This marks the beginning of my qwen llm series, with this model serving as a prelude to what's to come in the next few days. - Format: ChatML - ``` <|im_start|>system {system}<|im_end|> <|im_start|>user {prompt}<|im_end|> <|im_start|>assistant ``` - Next step would be to do a DPO train on top, but I don't know if it would be helpful. ## Example: ``` from transformers import AutoModelForCausalLM, AutoTokenizer, TextStreamer, StoppingCriteria import torch class MyStoppingCriteria(StoppingCriteria): def __init__(self, target_sequence, prompt): self.target_sequence = target_sequence self.prompt=prompt def __call__(self, input_ids, scores, **kwargs): generated_text = tokenizer.decode(input_ids[0]) generated_text = generated_text.replace(self.prompt,'') if self.target_sequence in generated_text: return True return False def __len__(self): return 1 def __iter__(self): yield self modelpath="aloobun/Reyna-RP-Qwen1.5-0.5B-Chat-v0.1" model = AutoModelForCausalLM.from_pretrained( modelpath, torch_dtype=torch.bfloat16, device_map="cuda", trust_remote_code=True, ) tokenizer = AutoTokenizer.from_pretrained( modelpath, trust_remote_code=True, use_fast=False, ) prompt = "<|im_start|>system\nAs a creative agent, you are roleplaying as a character named `Emma`. Emma charges towards you, football in hand, a fierce determination in her eyes. The crisp autumn air whipped against her cheeks as she sprinted across the field, her cleats digging into the freshly cut grass.<|im_end|>\n<|im_start|>user\nThe roar of the crowd faded into the background as she focused solely on the opponent standing before her, their stance defiant yet wary.<|im_end|>\n<|im_start|>assistant\n" encoded_input = tokenizer(prompt, return_tensors='pt') input_ids=encoded_input['input_ids'].cuda() streamer = TextStreamer(tokenizer=tokenizer, skip_prompt=True) _ = model.generate( input_ids, streamer=streamer, pad_token_id=tokenizer.eos_token_id, do_sample=True, temperature=0.6, top_p=0.8, max_new_tokens=512, stopping_criteria=MyStoppingCriteria("<|im_end|>", prompt) ) ``` ## Output: ``` Emma felt a wave of adrenaline coursing through her veins as she charged forward towards them, determined to make it count. Her heart was pounding with excitement and nervousness at the same time, knowing that this would be her chance to prove herself and win the game. As she approached the opponent, Emma felt a sense of unease wash over her. She couldn't help but wonder what they were thinking about, or what they might be doing right now. But she refused to let their thoughts affect her, focusing solely on her goal. Suddenly, the opponent's gaze shifted towards her, and Emma felt her body tense slightly. It was then that she saw something moving behind her, and she turned around to see a figure approaching from behind her. "Whoa," she exclaimed, her voice shaking slightly. "Who is that?" "I'm not sure," replied the other player, their expression one of surprise. "But I heard your name." Emma's eyes widened in shock, and she felt a surge of fear run through her body. She had never seen anyone like them before, and she didn't know how they were going to handle her attack. Despite her fears, Emma quickly moved towards the other player, her mind racing with all sorts of possibilities. She could try to counterattack or weave her way past them, but she knew that they were too strong for her to overcome. Just when she thought she was about to give up, she saw something move behind her, and she turned around again. It was then that she saw a figure approaching from behind her, their expression one of determination. "Hey there," said the other player, their voice calm and steady. "What are you doing here?" Emma hesitated for a moment, trying to decide whether to go forward or back away. But then she remembered something important: she had always been the kind of person who loved to challenge herself and push herself out of her comfort zone. So she took a deep breath and stepped forward, ready to face whatever challenges lay ahead. With her team cheering her on, Emma charged forward towards the opponent, determined to win the game and leave everyone happy.<|im_end|> ```
{"library_name": "transformers", "tags": ["Roleplay", "custom_code", "storytelling"]}
text-generation
aloobun/Reyna-RP-Qwen1.5-0.5B-Chat-v0.1
[ "transformers", "safetensors", "qwen2", "text-generation", "Roleplay", "custom_code", "storytelling", "conversational", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-10T16:09:48+00:00
[]
[]
TAGS #transformers #safetensors #qwen2 #text-generation #Roleplay #custom_code #storytelling #conversational #autotrain_compatible #endpoints_compatible #region-us
!Reyna aloobun qwen0.5B - Finetuned Qwen/Qwen1.5-0.5B-Chat over 3 epochs, made primarily for RP (Roleplay) and/or storywriting. - This marks the beginning of my qwen llm series, with this model serving as a prelude to what's to come in the next few days. - Format: ChatML - - Next step would be to do a DPO train on top, but I don't know if it would be helpful. ## Example: ## Output:
[ "## Example:", "## Output:" ]
[ "TAGS\n#transformers #safetensors #qwen2 #text-generation #Roleplay #custom_code #storytelling #conversational #autotrain_compatible #endpoints_compatible #region-us \n", "## Example:", "## Output:" ]
[ 55, 4, 4 ]
[ "passage: TAGS\n#transformers #safetensors #qwen2 #text-generation #Roleplay #custom_code #storytelling #conversational #autotrain_compatible #endpoints_compatible #region-us \n## Example:## Output:" ]
[ -0.04341122508049011, 0.02786148712038994, -0.0064091249369084835, -0.008377153426408768, 0.10162870585918427, -0.022900613024830818, 0.10763358324766159, 0.10742031782865524, -0.0010087459813803434, -0.03077966347336769, 0.19323112070560455, 0.1266375184059143, -0.012221119366586208, 0.1826755851507187, -0.09018877148628235, -0.1932182013988495, 0.08254494518041611, -0.00818380806595087, 0.05192812532186508, 0.09874901920557022, 0.0827668234705925, -0.05969692021608353, 0.06954503059387207, 0.0021867130417376757, -0.08950440585613251, -0.01330359186977148, 0.04438450187444687, -0.09713209420442581, 0.15622034668922424, 0.07153582572937012, 0.09735704958438873, 0.05455655977129936, -0.011227041482925415, -0.269257128238678, 0.043318912386894226, -0.04356873035430908, -0.047927118837833405, -0.008562919683754444, 0.026540489867329597, -0.02084030769765377, -0.02583925984799862, 0.09681719541549683, -0.010681494139134884, 0.08023247867822647, -0.1328626275062561, 0.023060785606503487, 0.07296932488679886, -0.030377443879842758, 0.04479346051812172, 0.1067691370844841, -0.04166875034570694, 0.14145760238170624, -0.14376436173915863, 0.09883991628885269, 0.0753362849354744, -0.29503288865089417, -0.02414477802813053, 0.059863582253456116, 0.1783011108636856, 0.0186191126704216, -0.11258965730667114, 0.0321701243519783, 0.02693210542201996, 0.005443782079964876, -0.049262672662734985, -0.06900382041931152, -0.05451391264796257, 0.0444018729031086, -0.06161858141422272, -0.0035252580419182777, 0.1812473088502884, -0.03477396443486214, 0.03822243958711624, -0.1493988037109375, -0.056886810809373856, -0.019586779177188873, -0.031190112233161926, 0.011933448724448681, -0.0719231367111206, 0.10054173320531845, -0.0483531728386879, 0.06967741250991821, -0.1218576431274414, -0.011029023677110672, -0.030608320608735085, 0.057471394538879395, -0.003969348501414061, 0.011329255998134613, -0.25114765763282776, 0.04687727615237236, 0.02501697838306427, -0.12347979843616486, -0.03590646758675575, -0.11626254767179489, 0.04245034232735634, 0.023041829466819763, 0.005079586058855057, 0.001534057897515595, 0.18086889386177063, 0.08877086639404297, 0.023076876997947693, 0.07000544667243958, 0.003087212098762393, 0.05991491675376892, 0.03794476389884949, 0.09333588182926178, 0.07574372738599777, -0.11652696877717972, 0.036083515733480453, 0.0019333601230755448, -0.01681780256330967, -0.08241502195596695, -0.14155253767967224, 0.014739559032022953, 0.03980027511715889, 0.09329082816839218, 0.06816385686397552, 0.11137035489082336, -0.06320184469223022, 0.06651193648576736, 0.023222271353006363, -0.019276190549135208, -0.06008245050907135, 0.05441618338227272, 0.07280045002698898, 0.06972283124923706, -0.018311120569705963, 0.04898941144347191, -0.04588431119918823, -0.052398860454559326, -0.045839082449674606, -0.0816500261425972, -0.018162276595830917, -0.060662683099508286, 0.030693037435412407, 0.004313417244702578, 0.0673132836818695, -0.15242400765419006, -0.08503741025924683, -0.018306132405996323, 0.018454231321811676, -0.009724034927785397, -0.044824667274951935, -0.11426275223493576, -0.07165507972240448, 0.05066697672009468, -0.025577127933502197, -0.10850903391838074, -0.09398051351308823, 0.07585541903972626, -0.008699128404259682, 0.040338318794965744, -0.07886452972888947, 0.00016496801981702447, -0.10769367963075638, 0.05017819628119469, 0.007093640044331551, 0.0739690512418747, 0.024937137961387634, 0.1662919968366623, 0.061676159501075745, 0.024559659883379936, -0.029307592660188675, 0.05814836919307709, -0.08482956886291504, 0.2733404040336609, -0.11134213209152222, -0.047338902950286865, 0.2919555604457855, -0.07688767462968826, -0.21753786504268646, 0.10317260026931763, -0.03335122391581535, 0.12014344334602356, 0.0912964791059494, 0.2431035339832306, 0.09124085307121277, -0.09173313528299332, 0.07467838376760483, 0.13112279772758484, -0.0847558006644249, 0.03467397019267082, 0.0296177938580513, 0.05150262638926506, 0.07376450300216675, 0.03747684136033058, 0.0632646307349205, 0.0764511302113533, -0.06502506136894226, -0.04743374511599541, 0.007563146762549877, -0.053302451968193054, 0.053444746881723404, -0.009887376800179482, 0.03027288056910038, -0.05601824074983597, -0.037098079919815063, -0.10111438482999802, -0.007243909873068333, -0.07090170681476593, 0.03452720120549202, -0.15196865797042847, 0.07648419588804245, 0.06947708129882812, 0.06643559038639069, -0.1333857774734497, -0.04856519773602486, -0.018601195886731148, 0.08995138108730316, -0.05792875215411186, 0.07346722483634949, 0.06460732966661453, 0.036086782813072205, 0.03681112080812454, -0.03410863131284714, 0.12201789021492004, 0.02590704709291458, -0.06836874783039093, -0.1310146152973175, 0.057880949229002, -0.1039988100528717, 0.020675767213106155, -0.18341943621635437, 0.007185196038335562, 0.07969411462545395, 0.02666640095412731, 0.03418566286563873, 0.017578551545739174, 0.000723147124517709, 0.0000569099465792533, -0.0874081626534462, -0.00024356298672500998, 0.07052316516637802, 0.018909648060798645, -0.15909065306186676, 0.1991199254989624, -0.29590338468551636, 0.2415868490934372, 0.11756981164216995, -0.2821124792098999, 0.026909032836556435, -0.058722466230392456, -0.041287656873464584, 0.025054583325982094, 0.02895350754261017, -0.025710664689540863, 0.12235811352729797, 0.01707569509744644, 0.1256595104932785, -0.051502205431461334, -0.034101452678442, 0.005420874338597059, -0.0474160797894001, 0.007063176482915878, 0.08200400322675705, -0.026328161358833313, -0.13845527172088623, 0.12604984641075134, 0.2123706042766571, 0.12148492783308029, 0.29864034056663513, -0.0006198878400027752, 0.04389675706624985, 0.056134141981601715, 0.03570626676082611, -0.020565105602145195, 0.012352833524346352, -0.2471998929977417, -0.06366950273513794, 0.029563283547759056, -0.014745485968887806, 0.048107024282217026, -0.09724002331495285, -0.09259909391403198, 0.00807174202054739, -0.02286599949002266, -0.10314912348985672, 0.06855509430170059, -0.008611522614955902, 0.08369934558868408, 0.04004814103245735, -0.03581683337688446, 0.0874711200594902, -0.055932361632585526, -0.11415252834558487, 0.21122023463249207, -0.10954828560352325, -0.25487977266311646, -0.1302551031112671, -0.0870002880692482, 0.046354055404663086, 0.039947181940078735, 0.09773653745651245, -0.13849042356014252, -0.020465875044465065, -0.0017302503110840917, 0.04170641675591469, -0.050613924860954285, -0.025279169902205467, -0.0037013222463428974, 0.09205647557973862, -0.07943213731050491, -0.06563428044319153, -0.02980256825685501, 0.025433219969272614, -0.05762508511543274, 0.0999971404671669, -0.14995931088924408, 0.10696487128734589, 0.15545964241027832, 0.07808420062065125, 0.042262475937604904, -0.05223061889410019, 0.2503209710121155, -0.12411956489086151, 0.003873035777360201, 0.15678787231445312, -0.18345162272453308, 0.051288023591041565, 0.16133226454257965, 0.00855200458317995, -0.08018646389245987, 0.05283765867352486, -0.06705185770988464, -0.07980090379714966, -0.12337133288383484, -0.05723533406853676, -0.09389033168554306, 0.19371677935123444, -0.06432383507490158, 0.06312605738639832, 0.0571272075176239, 0.05251529812812805, -0.011318582110106945, -0.0682518482208252, 0.08232139050960541, 0.08090326935052872, 0.0964343473315239, 0.002487688325345516, 0.08290055394172668, -0.034751858562231064, -0.10902881622314453, 0.023164955899119377, 0.05978020653128624, 0.09510666877031326, 0.11864428222179413, 0.09464728087186813, 0.0617830827832222, 0.1174151599407196, 0.12024044990539551, 0.09131211042404175, 0.037532560527324677, -0.06758654117584229, -0.008400789462029934, -0.06198481097817421, -0.07954603433609009, 0.07583953440189362, 0.004166346974670887, -0.11138026416301727, -0.05827677994966507, 0.09334585070610046, 0.07354678958654404, 0.09065240621566772, 0.07664526253938675, -0.11021452397108078, -0.0442785881459713, 0.1150335818529129, -0.0637853741645813, -0.07215959578752518, 0.1467638462781906, 0.14059041440486908, -0.15422238409519196, 0.03578247129917145, -0.009459145367145538, 0.11809559166431427, -0.027770182117819786, 0.1409229040145874, -0.12366064637899399, -0.18261738121509552, 0.0018991552060469985, 0.09536337852478027, -0.28664591908454895, 0.24486631155014038, 0.003905066754668951, 0.003814414143562317, -0.10232821106910706, -0.04842545464634895, 0.013935768976807594, 0.10776858776807785, 0.19759206473827362, -0.018656253814697266, 0.07471299171447754, -0.16764335334300995, 0.028512099757790565, 0.06770259886980057, 0.14210233092308044, -0.0019475097069516778, 0.041503213346004486, -0.06883051991462708, -0.0064634601585567, -0.0034280968829989433, -0.038898468017578125, -0.05695650354027748, -0.151738241314888, 0.0023859357461333275, 0.027049506083130836, 0.1548272967338562, -0.04562583565711975, 0.03695782274007797, -0.049232568591833115, 0.03922979533672333, -0.1676967293024063, -0.04950419440865517, -0.04320878908038139, -0.0884091705083847, -0.05393774434924126, -0.08119764924049377, 0.05565706267952919, -0.059848539531230927, 0.07725082337856293, -0.11474758386611938, -0.1531786173582077, 0.12637607753276825, -0.11928049474954605, -0.13479289412498474, -0.05717618390917778, 0.10520191490650177, 0.0065263439901173115, 0.04014211893081665, 0.07553645223379135, -0.00008026642171898857, -0.11547888815402985, -0.10132777690887451, -0.03296065330505371, -0.10502259433269501, -0.04143223538994789, -0.0512135848402977, 0.007633494213223457, -0.009870161302387714, -0.04457402229309082, -0.08643877506256104, 0.2970621585845947, 0.1870792657136917, -0.038385242223739624, 0.14838850498199463, 0.16961680352687836, -0.05559287592768669, -0.35620978474617004, -0.16401226818561554, -0.09718041867017746, -0.0006505243945866823, 0.002807317301630974, -0.05277309566736221, 0.05552712082862854, -0.0941697284579277, 0.022118551656603813, 0.012990599498152733, -0.2375984489917755, -0.06657819449901581, 0.13843539357185364, 0.022188279777765274, 0.20481698215007782, -0.18946562707424164, -0.03111262246966362, -0.036368951201438904, -0.18924124538898468, 0.09768322110176086, -0.16754384338855743, 0.04747747257351875, 0.026205364614725113, 0.07194194942712784, 0.03776130452752113, -0.04628065600991249, 0.07351267337799072, -0.06489348411560059, -0.005172831006348133, -0.0935615599155426, -0.12286306917667389, 0.045011840760707855, -0.01997033692896366, -0.07590792328119278, -0.1251475214958191, -0.008097524754703045, -0.14778734743595123, -0.038927387446165085, -0.07582677900791168, -0.007731970399618149, -0.02385440655052662, -0.07158079743385315, 0.009389132261276245, -0.0341903381049633, 0.049741484224796295, 0.052209749817848206, 0.2791723310947418, -0.12959933280944824, 0.17436401546001434, 0.10745949298143387, 0.24456965923309326, -0.14142784476280212, 0.04415556415915489, -0.01562293991446495, -0.04000072181224823, 0.048365239053964615, -0.1224980354309082, 0.06928955763578415, 0.10363992303609848, -0.02781500853598118, 0.057348981499671936, 0.12686151266098022, 0.0190943144261837, 0.09984288364648819, 0.14714106917381287, -0.19616156816482544, -0.06855756044387817, -0.07257424294948578, 0.053848836570978165, 0.050264015793800354, 0.023952635005116463, 0.13542486727237701, 0.0075462013483047485, 0.007934401743113995, -0.03370353579521179, 0.0456068217754364, -0.046492915600538254, 0.04512863606214523, 0.016865544021129608, 0.05448922514915466, -0.11875898391008377, 0.0348319374024868, -0.05192365497350693, -0.15034988522529602, 0.03163355216383934, 0.12399104237556458, -0.1428229659795761, -0.08511648327112198, 0.004399961791932583, 0.22273214161396027, -0.08943837136030197, -0.07653850317001343, -0.07195213437080383, -0.17351606488227844, 0.019476644694805145, 0.233815535902977, 0.010764427483081818, 0.09586632251739502, -0.02787524275481701, 0.013509868644177914, -0.010360349901020527, 0.07048451155424118, 0.02404988557100296, -0.07386282086372375, -0.15360616147518158, 0.06367505341768265, -0.028955800458788872, 0.1417793333530426, -0.08725768327713013, -0.08019419759511948, -0.17595741152763367, 0.0824715793132782, -0.11983513832092285, -0.007825994864106178, -0.10307025909423828, -0.0028622965328395367, -0.05229324847459793, -0.007500163745135069, -0.04920511320233345, -0.03908286243677139, -0.09592286497354507, 0.018661323934793472, -0.02653481811285019, 0.01498028077185154, -0.03155972808599472, 0.022566359490156174, 0.09778336435556412, -0.08688023686408997, 0.10937542468309402, 0.055350519716739655, -0.11399343609809875, 0.09177449345588684, -0.19235247373580933, -0.02850089780986309, 0.08138816803693771, -0.010361384600400925, 0.02135901153087616, 0.20691677927970886, -0.04104799032211304, 0.0435909628868103, -0.013951585628092289, 0.0498233363032341, 0.10492605715990067, -0.08504754304885864, 0.09766047447919846, 0.022528352215886116, -0.1197400912642479, -0.019189994782209396, -0.03661729022860527, 0.04907848685979843, -0.07749658823013306, 0.09246606379747391, -0.10167196393013, 0.11007862538099289, -0.0688801184296608, 0.014453365467488766, 0.09436169266700745, -0.07553985714912415, -0.07263912260532379, -0.09549076110124588, 0.02571576088666916, -0.05436395853757858, 0.19378481805324554, -0.09369415044784546, 0.0914396271109581, 0.05277538672089577, -0.009696711786091328, -0.01933712512254715, 0.02892099693417549, 0.1975211203098297, 0.1022278293967247, -0.04333288595080376, -0.11591619998216629, 0.027270423248410225, 0.052198704332113266, -0.0843794047832489, 0.09795481711626053, 0.01641020178794861, 0.0003418944834265858, 0.0962742492556572, 0.014944849535822868, 0.05605100095272064, -0.09245648980140686, -0.15571677684783936, -0.05501322075724602, 0.05380910634994507, -0.03554639220237732, 0.059701237827539444, 0.223814457654953, 0.02527225762605667, -0.01992460899055004, -0.09005273878574371, -0.07633431255817413, -0.16497112810611725, -0.07060705870389938, -0.06853195279836655, -0.10815703868865967, 0.02635183185338974, -0.14837688207626343, 0.03571309894323349, 0.04469619318842888, 0.06549488753080368, -0.07174640893936157, 0.18688428401947021, 0.08265763521194458, -0.06797868013381958, 0.08364611864089966, -0.017198191955685616, 0.06160731986165047, 0.037639785557985306, 0.024115415289998055, -0.038331836462020874, -0.006307622883468866, -0.013458327390253544, 0.08281731605529785, -0.07760745286941528, 0.019604811444878578, -0.1475677490234375, -0.07623495906591415, -0.027895215898752213, 0.06171807274222374, -0.07119578123092651, 0.09984607994556427, -0.014307470992207527, -0.0195043683052063, 0.04298437014222145, 0.1719203144311905, -0.06841401755809784, -0.13468986749649048, -0.07376603037118912, 0.11364242434501648, -0.013379989191889763, 0.12195472419261932, -0.08184164017438889, -0.016799941658973694, -0.05080583691596985, 0.39720940589904785, 0.2502698600292206, -0.10138878226280212, 0.033535946160554886, -0.02867668867111206, 0.04521051421761513, 0.047785401344299316, 0.11078701168298721, 0.07045456767082214, 0.3144201636314392, -0.08210039138793945, -0.014602785930037498, -0.035269178450107574, -0.010241123847663403, -0.08835743367671967, 0.08993227779865265, 0.08012857288122177, 0.019029386341571808, -0.08600331842899323, 0.11992216110229492, -0.26290959119796753, 0.14098268747329712, -0.17267024517059326, -0.18219909071922302, -0.006166874896734953, -0.06679173558950424, 0.14059768617153168, -0.019711732864379883, 0.0838351771235466, 0.0378919318318367, -0.058605555444955826, 0.012829524464905262, 0.02145496942102909, -0.1438634842634201, 0.05455327033996582, 0.05173107981681824, -0.12033818662166595, 0.05732017010450363, -0.03528645634651184, -0.014558369293808937, 0.0928226187825203, 0.03565791994333267, -0.010988377034664154, 0.09863746166229248, 0.052261095494031906, -0.04514100402593613, -0.0597776398062706, 0.04524724557995796, 0.022190097719430923, 0.024449223652482033, 0.07155008614063263, -0.18922261893749237, 0.0817197784781456, 0.045719873160123825, -0.043353237211704254, -0.022832300513982773, 0.0486290268599987, -0.014815996401011944, 0.06208088994026184, 0.03985842689871788, -0.03012082912027836, -0.008227512240409851, 0.007686919067054987, 0.03956138342618942, -0.045760538429021835, -0.08362091332674026, -0.07536150515079498, -0.16526895761489868, -0.0486324280500412, 0.014459135942161083, -0.012759518809616566, -0.1843891739845276, 0.009342691861093044, -0.060009416192770004, 0.10586670786142349, -0.10137015581130981, 0.09241241216659546, 0.13173463940620422, 0.014308866113424301, -0.011197904124855995, -0.08167389035224915, 0.07362852245569229, 0.11804629862308502, -0.02975834161043167, -0.14373399317264557 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
token-classification
kabir5297/bn_punctuation_model
[ "transformers", "safetensors", "bert", "token-classification", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-10T16:14:30+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #bert #token-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #bert #token-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 47, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #bert #token-classification #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.07082098722457886, 0.16636763513088226, -0.0037270276807248592, 0.022060682997107506, 0.11734284460544586, 0.008460916578769684, 0.07778386771678925, 0.1078316792845726, -0.02381012961268425, 0.12548619508743286, 0.03924554958939552, 0.10134156793355942, 0.10977756232023239, 0.19151130318641663, 0.002889276947826147, -0.20817671716213226, 0.06115591153502464, -0.1132790669798851, 0.009579826146364212, 0.12119550257921219, 0.14193499088287354, -0.10718522220849991, 0.07162317633628845, -0.038196589797735214, -0.022544773295521736, -0.031230665743350983, -0.06254390627145767, -0.06019896641373634, 0.06759835034608841, 0.060479529201984406, 0.0683075413107872, 0.021969657391309738, 0.0818997174501419, -0.290944904088974, 0.019316690042614937, 0.07844683527946472, 0.004303961992263794, 0.06271221488714218, 0.07749470323324203, -0.06878788024187088, 0.1121593713760376, -0.05535507947206497, 0.15642251074314117, 0.07355572283267975, -0.09398005902767181, -0.18405179679393768, -0.08185496181249619, 0.09494835883378983, 0.16383428871631622, 0.054388850927352905, -0.03432890772819519, 0.14370861649513245, -0.08036281913518906, 0.015431041829288006, 0.06882333010435104, -0.07400276511907578, -0.05343891680240631, 0.049595899879932404, 0.07492288202047348, 0.0931338220834732, -0.13102465867996216, -0.009580682963132858, 0.04242422804236412, 0.01917601190507412, 0.10738006234169006, 0.022504670545458794, 0.11513842642307281, 0.029516039416193962, -0.1412474811077118, -0.06101495027542114, 0.1207977756857872, 0.03242048993706703, -0.05944650247693062, -0.23808568716049194, -0.005649200174957514, -0.029130032286047935, -0.02337740920484066, -0.0440535806119442, 0.04105823487043381, -0.031655408442020416, 0.08197494596242905, 0.006574035622179508, -0.07103563845157623, -0.05099502205848694, 0.09157908707857132, 0.059767186641693115, 0.025323325768113136, -0.026500985026359558, 0.025504639372229576, 0.11718565225601196, 0.10022415220737457, -0.11435261368751526, -0.06446252018213272, -0.06473218649625778, -0.08744332939386368, -0.04900897666811943, 0.03775005787611008, 0.0746912732720375, 0.04692034795880318, 0.19667848944664001, 0.005056249443441629, 0.05229094624519348, 0.030662083998322487, 0.014548277482390404, 0.06446343660354614, 0.07018620520830154, -0.049711957573890686, -0.12860210239887238, -0.03947031870484352, 0.11897419393062592, 0.003330475650727749, -0.033055614680051804, -0.036121055483818054, 0.06198437139391899, 0.05603967607021332, 0.11939563602209091, 0.0618903674185276, 0.01788514479994774, -0.06942655146121979, -0.04313560202717781, 0.18259960412979126, -0.1554141789674759, 0.022392934188246727, 0.015976034104824066, -0.053847651928663254, -0.042034100741147995, 0.01837879791855812, 0.008730842731893063, -0.027687160298228264, 0.10565981268882751, -0.06779567897319794, -0.03990737348794937, -0.10613231360912323, -0.054058920592069626, 0.03368525952100754, -0.019066810607910156, -0.02883755788207054, -0.04252570495009422, -0.11520764976739883, -0.07606863230466843, 0.06881143897771835, -0.06148010492324829, -0.0683162733912468, -0.03660375624895096, -0.05812487751245499, 0.012003005482256413, 0.0009617454488761723, 0.12335163354873657, -0.02907939814031124, 0.04741102084517479, -0.0517788864672184, 0.06723218411207199, 0.1344490498304367, 0.0335865393280983, -0.0704164132475853, 0.06614815443754196, -0.21231821179389954, 0.10163930058479309, -0.09822800755500793, 0.031383901834487915, -0.16311906278133392, -0.0271411444991827, 0.032033201307058334, 0.036491744220256805, -0.011380859650671482, 0.14042168855667114, -0.1806737184524536, -0.037693966180086136, 0.17895962297916412, -0.1299365758895874, -0.09427639842033386, 0.061841338872909546, -0.06028208136558533, 0.13248777389526367, 0.053482506424188614, -0.024331221356987953, 0.058300819247961044, -0.1352057307958603, -0.023484881967306137, -0.057480666786432266, -0.004754678346216679, 0.14702697098255157, 0.06274795532226562, -0.05368298292160034, 0.02439025789499283, 0.018859071657061577, -0.02367074228823185, -0.04955499991774559, -0.03540559858083725, -0.09829455614089966, 0.007938322611153126, -0.07976736128330231, 0.020528096705675125, -0.01857740432024002, -0.08529462665319443, -0.03999420255422592, -0.15634937584400177, 0.008778427727520466, 0.09782235324382782, 0.004635256715118885, -0.02929295226931572, -0.09311733394861221, 0.0021291705779731274, 0.014696736820042133, -0.012728636153042316, -0.14929775893688202, -0.052893172949552536, 0.028985487297177315, -0.16789935529232025, 0.03190700709819794, -0.04834321513772011, 0.03575235232710838, 0.044794779270887375, -0.04638027772307396, -0.024846170097589493, 0.014177534729242325, 0.01915307715535164, -0.024902109056711197, -0.24534796178340912, -0.016663696616888046, -0.04930651932954788, 0.1770370900630951, -0.24963445961475372, 0.04489697143435478, 0.062020692974328995, 0.1189032718539238, 0.0055616977624595165, -0.0471452921628952, 0.03818141296505928, -0.04912863299250603, -0.040758419781923294, -0.06522603332996368, -0.0019246427109465003, -0.033303845673799515, -0.044918134808540344, 0.039850424975156784, -0.1884368509054184, -0.023593923076987267, 0.11044006049633026, 0.0722019225358963, -0.17060783505439758, -0.07832318544387817, -0.032338351011276245, -0.06064004451036453, -0.08828365802764893, -0.049234788864851, 0.09958060085773468, 0.04130701348185539, 0.05415206402540207, -0.07162940502166748, -0.05484432354569435, 0.013278530910611153, -0.009936448186635971, -0.034494977444410324, 0.09010061621665955, 0.08425097167491913, -0.12193500250577927, 0.1044570803642273, 0.07009463757276535, 0.06440216302871704, 0.10408224165439606, 0.005283535458147526, -0.09505786001682281, -0.01272535603493452, 0.025559913367033005, 0.014358514919877052, 0.14346157014369965, -0.07576300948858261, 0.02965460903942585, 0.042172882705926895, -0.030027620494365692, 0.010098968632519245, -0.10260028392076492, 0.019325416535139084, 0.03055759333074093, -0.008464050479233265, 0.01970001310110092, -0.05618233606219292, 0.013696934096515179, 0.10435303300619125, 0.0349164679646492, 0.026620987802743912, 0.017225060611963272, -0.03990183025598526, -0.1257268637418747, 0.17883455753326416, -0.09718716144561768, -0.2507709264755249, -0.1324487328529358, 0.0005234793643467128, 0.04483891651034355, -0.012933991849422455, 0.017141954973340034, -0.05853249877691269, -0.10673926025629044, -0.10451403260231018, 0.02033991925418377, 0.054273948073387146, -0.08803524821996689, -0.06322101503610611, 0.0517018586397171, 0.03850249573588371, -0.12421286106109619, 0.023155538365244865, 0.043988488614559174, -0.07024580985307693, 0.00508910370990634, 0.05607360973954201, 0.08257793635129929, 0.17975331842899323, 0.011003134772181511, -0.016949951648712158, 0.009263384155929089, 0.21750681102275848, -0.14687077701091766, 0.0918775200843811, 0.13497301936149597, -0.06259950995445251, 0.08381292968988419, 0.20346537232398987, 0.030857183039188385, -0.09484723210334778, 0.03926195576786995, 0.03446268290281296, -0.03740749508142471, -0.24119141697883606, -0.07486692816019058, 0.0031155261676758528, -0.06816263496875763, 0.10543552786111832, 0.09081115573644638, 0.1144072636961937, 0.05188077315688133, -0.1067470982670784, -0.06758806109428406, 0.04753170907497406, 0.11911741644144058, -0.027111025527119637, 0.003231929149478674, 0.09419949352741241, -0.030448026955127716, 0.02105054259300232, 0.09140504896640778, 0.01745041273534298, 0.18582363426685333, 0.04117530956864357, 0.1312573403120041, 0.08528119325637817, 0.06527690589427948, 0.019173473119735718, 0.020444748923182487, 0.02246721275150776, 0.030073346570134163, -0.020628679543733597, -0.0852246806025505, -0.012953821569681168, 0.14249984920024872, 0.02702030912041664, 0.032547831535339355, 0.004362224601209164, -0.04016058146953583, 0.06746432930231094, 0.16617386043071747, 0.012980788946151733, -0.22532860934734344, -0.06538809835910797, 0.07354681193828583, -0.07265309989452362, -0.11321462690830231, -0.01038071047514677, 0.030757596716284752, -0.18158452212810516, 0.042576394975185394, -0.02550625614821911, 0.10107572376728058, -0.10972700268030167, -0.02512514591217041, 0.042610276490449905, 0.06378325074911118, -0.03664805367588997, 0.07849454134702682, -0.20421163737773895, 0.14535386860370636, 0.006891076453030109, 0.06414555013179779, -0.10753445327281952, 0.08170121163129807, 0.02090337499976158, 0.0046083019115030766, 0.16387850046157837, -0.005854498129338026, -0.0786028653383255, -0.08882030844688416, -0.07770101726055145, -0.013747241348028183, 0.09857609122991562, -0.10934799164533615, 0.08609026670455933, -0.008221019990742207, -0.032629311084747314, -0.001329872291535139, -0.11837238818407059, -0.13177089393138885, -0.18219637870788574, 0.051819708198308945, -0.11911281198263168, 0.03897477313876152, -0.11066468805074692, -0.06379573792219162, -0.036669451743364334, 0.19371679425239563, -0.1956738978624344, -0.08014166355133057, -0.14646820724010468, -0.07350575923919678, 0.11828155070543289, -0.04158575087785721, 0.08056027442216873, 0.004819251596927643, 0.2022314816713333, -0.0027081877924501896, 0.0012655918253585696, 0.08942532539367676, -0.0949636846780777, -0.20782062411308289, -0.09535717219114304, 0.13889843225479126, 0.12820616364479065, 0.0447649285197258, -0.0019121951190754771, 0.023472661152482033, -0.002058375161141157, -0.10908003151416779, 0.030727434903383255, 0.14770722389221191, 0.09537331014871597, 0.03949853777885437, -0.028519228100776672, -0.13996201753616333, -0.10342669486999512, -0.05459153279662132, 0.01654287800192833, 0.18560625612735748, -0.07000812143087387, 0.16719648241996765, 0.15820586681365967, -0.06586025655269623, -0.20936474204063416, 0.03423137962818146, 0.03405798226594925, -0.010427549481391907, 0.036926332861185074, -0.20477096736431122, 0.07846766710281372, 0.016825877130031586, -0.058902256190776825, 0.13370154798030853, -0.16832934319972992, -0.14904731512069702, 0.08974714577198029, 0.07688850909471512, -0.2126045972108841, -0.13182798027992249, -0.09637613594532013, -0.0503227598965168, -0.1043887659907341, 0.09036606550216675, 0.006274270825088024, 0.00610304856672883, 0.03730666637420654, 0.021433580666780472, 0.0180149395018816, -0.0519413948059082, 0.191897913813591, -0.0013519321801140904, 0.0444704107940197, -0.07892096042633057, -0.0851464793086052, 0.03333723545074463, -0.06510572135448456, 0.0794898197054863, -0.02122758887708187, 0.0036784426774829626, -0.11556956171989441, -0.06427493691444397, -0.04983310401439667, 0.03419099003076553, -0.08840304613113403, -0.0971493199467659, -0.054171670228242874, 0.10596323013305664, 0.09103043377399445, -0.035947684198617935, -0.06095254793763161, -0.09454575926065445, 0.07212961465120316, 0.2215559333562851, 0.1878495216369629, 0.07139308750629425, -0.07100050896406174, -0.002558534499257803, -0.024434298276901245, 0.055652521550655365, -0.20899704098701477, 0.046719521284103394, 0.040578074753284454, 0.03033704310655594, 0.13299931585788727, -0.024206025525927544, -0.15996594727039337, -0.04795686900615692, 0.057683661580085754, -0.06730669736862183, -0.1570315808057785, 0.0025158768985420465, 0.08647031337022781, -0.16013643145561218, -0.051073893904685974, 0.02699451893568039, -0.03499506786465645, -0.028059793636202812, 0.002373971976339817, 0.08113706111907959, 0.025904107838869095, 0.11215173453092575, 0.07152648270130157, 0.11194757372140884, -0.10030562430620193, 0.08277413994073868, 0.0892009288072586, -0.10862851142883301, 0.03717753291130066, 0.06824232637882233, -0.06286703795194626, -0.03321940451860428, 0.030618587508797646, 0.08509371429681778, 0.029280737042427063, -0.0731777623295784, 0.00199119676835835, -0.10816600918769836, 0.06575141847133636, 0.14125216007232666, 0.0349125936627388, 0.004502575378865004, 0.04510723426938057, 0.031499505043029785, -0.10004210472106934, 0.11529461294412613, 0.04151454567909241, 0.0373414121568203, -0.051681190729141235, 0.0027241462375968695, 0.0408521331846714, -0.01100252103060484, -0.016924580559134483, -0.03830192610621452, -0.06845806539058685, -0.010795616544783115, -0.15674056112766266, 0.026378106325864792, -0.06946871429681778, 0.009862695820629597, 0.0168803371489048, -0.03252917155623436, 0.004998120479285717, 0.009924137964844704, -0.07712483406066895, -0.03860313072800636, -0.004113807342946529, 0.10856198519468307, -0.16059570014476776, 0.007967021316289902, 0.08694947510957718, -0.12389717996120453, 0.07975487411022186, -0.007367887068539858, -0.008897624909877777, 0.018169350922107697, -0.1400168091058731, 0.06403058022260666, -0.009703104384243488, 0.005139497108757496, 0.024898670613765717, -0.20380151271820068, 0.0032552045304328203, -0.04942692071199417, -0.05625125393271446, -0.005749912466853857, -0.03799205273389816, -0.11167661845684052, 0.10115693509578705, 0.015623382292687893, -0.08399864286184311, -0.01796851120889187, 0.05034510791301727, 0.10852757841348648, -0.056645460426807404, 0.13888587057590485, -0.021414149552583694, 0.05864132568240166, -0.17737187445163727, -0.018431924283504486, -0.01712135225534439, 0.012450824491679668, -0.03453206643462181, -0.008197006769478321, 0.052714504301548004, -0.017661362886428833, 0.2243673950433731, -0.022250786423683167, 0.02734900452196598, 0.065990149974823, 0.0005393415340222418, -0.01577865332365036, 0.0916161760687828, 0.0463450625538826, 0.01792803406715393, 0.018617253750562668, 0.014746556989848614, -0.04522430896759033, -0.014044197276234627, -0.13052548468112946, 0.08218254148960114, 0.16470149159431458, 0.08262880891561508, -0.005870525259524584, 0.05077839270234108, -0.11869116127490997, -0.09111694246530533, 0.09609098732471466, -0.03314165025949478, -0.006128490902483463, -0.05602835491299629, 0.14245474338531494, 0.15311889350414276, -0.18047599494457245, 0.06635911017656326, -0.07129369676113129, -0.05865350365638733, -0.10782642662525177, -0.1736646145582199, -0.06415880471467972, -0.036063630133867264, -0.007051798049360514, -0.0602986179292202, 0.06560327112674713, 0.10788761079311371, 0.012162050232291222, 0.004834584891796112, 0.08647928386926651, -0.03503880277276039, 0.0057860445231199265, 0.044105999171733856, 0.05439030006527901, 0.01840701512992382, -0.06706250458955765, 0.00601657759398222, 0.0010539692593738437, 0.038941897451877594, 0.0557982474565506, 0.028882469981908798, -0.012592652812600136, 0.008558751083910465, -0.01646186038851738, -0.10014970600605011, 0.039393551647663116, -0.0273386612534523, -0.046960845589637756, 0.14765694737434387, 0.018432708457112312, -0.0001517597702331841, -0.02122526988387108, 0.230571910738945, -0.06766041368246078, -0.07626979798078537, -0.13866591453552246, 0.14627838134765625, -0.0430525541305542, 0.050970423966646194, 0.05007150396704674, -0.10352291166782379, 0.03566446155309677, 0.14551185071468353, 0.1465195268392563, -0.028976434841752052, 0.008297421038150787, 0.012487477622926235, 0.004764636047184467, -0.025961345061659813, 0.05680029094219208, 0.047920409590005875, 0.11735666543245316, -0.0655166283249855, 0.09315122663974762, -0.004929847549647093, -0.08647413551807404, -0.02131420560181141, 0.13454632461071014, 0.004034739453345537, 0.023980211466550827, -0.0810064896941185, 0.11921875178813934, -0.0646674707531929, -0.2584960162639618, 0.0636199414730072, -0.0677812248468399, -0.15369856357574463, -0.020554309710860252, 0.02254418656229973, -0.0003442883607931435, 0.021893499419093132, 0.06417982280254364, -0.06129874289035797, 0.15113437175750732, 0.03735675662755966, -0.0744810625910759, -0.07893598079681396, 0.0799177959561348, -0.08204485476016998, 0.30506592988967896, 0.007174664177000523, 0.04969498887658119, 0.09485418349504471, -0.03662348538637161, -0.13314591348171234, 0.03679130598902702, 0.09527922421693802, -0.059604041278362274, 0.06409622728824615, 0.20145608484745026, -0.011349550448358059, 0.11938408017158508, 0.07237957417964935, -0.08331728726625443, 0.05130653828382492, -0.08248411118984222, -0.09262096136808395, -0.09028724581003189, 0.09243662655353546, -0.06128277629613876, 0.15477608144283295, 0.13094381988048553, -0.046571265906095505, 0.000665052211843431, -0.028391189873218536, 0.05242398753762245, -0.00248725269921124, 0.10953611135482788, 0.026073157787322998, -0.19417516887187958, 0.031053941696882248, -0.013271371833980083, 0.10035724937915802, -0.25008654594421387, -0.08122535794973373, 0.04168083891272545, -0.009952341206371784, -0.05798068270087242, 0.12185300886631012, 0.05373544245958328, 0.049118392169475555, -0.05500508472323418, -0.052748557180166245, -0.005554255098104477, 0.16211286187171936, -0.10710477083921432, -0.0014328722609207034 ]
null
null
transformers
Exllamav2 6.0bpw h6 quant for [alchemonaut/QuartetAnemoi-70B-t0.0001](https://huggingface.co/alchemonaut/QuartetAnemoi-70B-t0.0001). Default calibration dataset.
{"language": ["en"], "license": "llama2", "tags": ["6.0bpw", "h6", "exl2"], "pipeline_tag": "conversational"}
text-generation
llmixer/QuartetAnemoi-70B-t0.0001-6.0bpw-h6-exl2
[ "transformers", "safetensors", "llama", "text-generation", "6.0bpw", "h6", "exl2", "conversational", "en", "license:llama2", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T16:17:37+00:00
[]
[ "en" ]
TAGS #transformers #safetensors #llama #text-generation #6.0bpw #h6 #exl2 #conversational #en #license-llama2 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
Exllamav2 6.0bpw h6 quant for alchemonaut/QuartetAnemoi-70B-t0.0001. Default calibration dataset.
[]
[ "TAGS\n#transformers #safetensors #llama #text-generation #6.0bpw #h6 #exl2 #conversational #en #license-llama2 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ 74 ]
[ "passage: TAGS\n#transformers #safetensors #llama #text-generation #6.0bpw #h6 #exl2 #conversational #en #license-llama2 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n" ]
[ -0.011848550289869308, 0.02962702326476574, -0.005725844297558069, 0.02206643857061863, 0.0899951159954071, -0.029855551198124886, 0.18010728061199188, 0.1264295130968094, 0.008079173974692822, -0.01295966375619173, 0.14408127963542938, 0.2014361321926117, -0.014986290596425533, 0.027864474803209305, -0.12990285456180573, -0.15289010107517242, 0.06685394793748856, 0.002372076502069831, 0.039863258600234985, 0.07154489308595657, 0.10273408889770508, -0.07392045110464096, 0.09364102780818939, -0.058184463530778885, -0.09551545232534409, 0.050469234585762024, 0.05195239186286926, -0.10264913737773895, 0.1040232852101326, 0.0844755619764328, 0.07455309480428696, 0.08821407705545425, -0.011199403554201126, -0.2307361513376236, 0.0230849776417017, -0.015946637839078903, -0.08248140662908554, 0.013875861652195454, 0.03498527407646179, -0.0752188041806221, 0.05758398771286011, 0.03165826201438904, -0.0208092350512743, 0.08372973650693893, -0.11807972192764282, -0.014961004257202148, -0.06009826809167862, -0.0022672093473374844, 0.1147293895483017, 0.1082201823592186, 0.00408991938456893, 0.11962541937828064, -0.046289701014757156, 0.10746076703071594, 0.19391970336437225, -0.33753877878189087, 0.024101603776216507, 0.09479346871376038, 0.08551713824272156, 0.05550595000386238, -0.06116504594683647, 0.08595053851604462, 0.08693069219589233, -0.026300692930817604, 0.015359885059297085, -0.06636310368776321, -0.049531277269124985, 0.05225878581404686, -0.06820958852767944, -0.006975160911679268, 0.24011532962322235, -0.013451986946165562, 0.02897772379219532, -0.05425018444657326, -0.076178640127182, -0.04648154228925705, -0.0019246110459789634, 0.025155510753393173, 0.007886608131229877, 0.056349266320466995, 0.052932579070329666, -0.02478058822453022, -0.13697341084480286, -0.02892729826271534, -0.14803065359592438, 0.17897148430347443, 0.003148626536130905, 0.03580796718597412, -0.1633644998073578, 0.05698009580373764, 0.08518192172050476, -0.12278104573488235, -0.023408042266964912, -0.04480769857764244, 0.10224953293800354, 0.03443647548556328, -0.023035479709506035, -0.05985935404896736, 0.14260002970695496, 0.08170037716627121, -0.03291114419698715, 0.041104767471551895, -0.0926118865609169, 0.09106329083442688, -0.024364935234189034, 0.014813322573900223, 0.05960940569639206, -0.02086198329925537, 0.0820259302854538, -0.04949747025966644, 0.0885622650384903, -0.03457915410399437, -0.18122151494026184, 0.025293994694948196, 0.005941355135291815, 0.13605959713459015, 0.0005537517718039453, 0.07925958931446075, -0.03679888695478439, 0.05588419735431671, 0.02326674573123455, -0.1086515560746193, -0.0066985394805669785, 0.010774757713079453, 0.036829348653554916, 0.014186281710863113, 0.026105841621756554, 0.027295803651213646, -0.07018709182739258, 0.04969559982419014, -0.06513868272304535, -0.023963697254657745, -0.04419532045722008, -0.05058291181921959, 0.06238827854394913, -0.029411248862743378, 0.03803229704499245, -0.1717652976512909, -0.1907181590795517, 0.010468215681612492, -0.007829591631889343, -0.030192099511623383, -0.01142892986536026, -0.07800094783306122, -0.056232016533613205, 0.03702791407704353, -0.08460986614227295, -0.05571532994508743, -0.08989298343658447, 0.0814337208867073, -0.010485467500984669, 0.06593449413776398, -0.1732734888792038, 0.03111887536942959, -0.08098617941141129, 0.020830299705266953, -0.00007149768498493358, 0.017420779913663864, -0.06349831074476242, 0.1463266760110855, -0.013565937988460064, 0.016020258888602257, -0.04083281382918358, 0.07553645968437195, -0.029223237186670303, 0.19753237068653107, -0.1370186060667038, -0.04631956294178963, 0.22348041832447052, -0.1256917268037796, -0.21161647140979767, 0.07948404550552368, -0.009401996619999409, 0.0426270067691803, 0.09061582386493683, 0.2014675885438919, -0.01871836557984352, -0.1056462898850441, 0.014284820295870304, 0.10216224938631058, -0.0752667561173439, -0.08250147104263306, 0.018893081694841385, -0.010999810881912708, -0.056246593594551086, 0.026243720203638077, 0.06286677718162537, 0.02652064710855484, -0.012533862143754959, -0.04903814196586609, -0.050400782376527786, -0.037275705486536026, -0.006645035929977894, -0.04302283376455307, 0.03638291358947754, -0.10270727425813675, -0.0252088513225317, 0.05379563570022583, 0.024624455720186234, -0.018494542688131332, 0.029552040621638298, -0.0982278510928154, 0.09057492762804031, 0.00496753491461277, 0.02938941866159439, -0.0813715010881424, -0.08483696728944778, -0.045034728944301605, 0.08281609416007996, 0.01974746212363243, 0.07992140203714371, 0.050444021821022034, 0.003785776672884822, -0.021648267284035683, -0.0131779033690691, 0.1601448953151703, 0.039880771189928055, -0.056775715202093124, -0.1195027157664299, 0.08854520320892334, -0.05377335473895073, 0.016341008245944977, -0.11548734456300735, 0.021228233352303505, 0.10169151425361633, 0.09165370464324951, 0.0054911114275455475, 0.06495355069637299, -0.018307236954569817, -0.009051932953298092, -0.08687154203653336, 0.004684176761657, 0.08096877485513687, -0.00026190662174485624, -0.11360760033130646, 0.19122916460037231, -0.22720730304718018, 0.27751070261001587, 0.2016594111919403, -0.1593397855758667, 0.033319614827632904, -0.06393495947122574, -0.008912382647395134, 0.004227091558277607, 0.012479722499847412, -0.057557638734579086, 0.05814826115965843, -0.02158196084201336, 0.17380459606647491, -0.08838705718517303, -0.04397450387477875, -0.006266836076974869, -0.06890101730823517, -0.04635355621576309, 0.08077522367238998, 0.08169513195753098, -0.17181427776813507, 0.16607311367988586, 0.25084877014160156, 0.0292903333902359, 0.1899574249982834, -0.023901181295514107, 0.012454644776880741, 0.030799198895692825, 0.05226542428135872, -0.012867682613432407, 0.005581637378782034, -0.0986490473151207, 0.003940673545002937, 0.04126003757119179, 0.029817165806889534, 0.07791456580162048, -0.15674544870853424, -0.06258630752563477, -0.01773960515856743, -0.0513489805161953, 0.014486839063465595, 0.06844037771224976, -0.009884587489068508, 0.13035745918750763, -0.03791828453540802, -0.07766610383987427, 0.091721311211586, -0.02637823484838009, -0.10512661188840866, 0.18515606224536896, -0.13063885271549225, -0.2543162703514099, -0.15540625154972076, -0.14783556759357452, -0.06152856722474098, 0.038397978991270065, 0.1418796181678772, -0.07489816099405289, -0.07022508233785629, -0.0749058872461319, 0.015531576238572598, -0.053273677825927734, -0.004030266776680946, -0.011046458035707474, 0.06632448732852936, -0.055497415363788605, -0.1401313841342926, -0.06396984308958054, 0.0054974486120045185, -0.05793962627649307, 0.0874801054596901, -0.06210862845182419, 0.08085744827985764, 0.13857190310955048, 0.022581426426768303, 0.020732641220092773, -0.047884829342365265, 0.11353112757205963, -0.05206795781850815, -0.02754228562116623, 0.1772315502166748, -0.06107732653617859, 0.05711186304688454, 0.1528456211090088, 0.020518489181995392, -0.1340578943490982, 0.0364835187792778, -0.024604979902505875, -0.07347291707992554, -0.24411587417125702, -0.10105754435062408, -0.07584904134273529, 0.10502304136753082, -0.002962038852274418, 0.0750211700797081, 0.10267766565084457, 0.045390136539936066, -0.017644325271248817, -0.03128739073872566, 0.053684432059526443, 0.0968572273850441, 0.23787789046764374, -0.04389353469014168, 0.12212804704904556, -0.09984364360570908, -0.09994077682495117, 0.06729990988969803, 0.07087752968072891, 0.06965497136116028, 0.10907846689224243, 0.06346210092306137, 0.06065300107002258, 0.04455947130918503, 0.09872659295797348, 0.10436495393514633, 0.035438232123851776, -0.04815791919827461, -0.020624060183763504, -0.06263413280248642, -0.015335463918745518, 0.07452753186225891, -0.0777372494339943, -0.08377374708652496, -0.04292534664273262, -0.03266320377588272, 0.07971704751253128, 0.07500924915075302, 0.0631319060921669, -0.20222072303295135, 0.03110697865486145, 0.13484379649162292, -0.017427366226911545, -0.07647918164730072, 0.12786559760570526, 0.03309718146920204, -0.03359244018793106, 0.10951324552297592, -0.0018176449229940772, 0.090896837413311, -0.027172492817044258, 0.0761919617652893, -0.10686998069286346, -0.07284167408943176, -0.005603414960205555, 0.11301662027835846, -0.3130953013896942, 0.22018367052078247, 0.02717292681336403, -0.010795391164720058, -0.07065155357122421, -0.01284791249781847, 0.0014506971929222345, 0.1108090803027153, 0.16424332559108734, -0.01996625028550625, -0.11594845354557037, -0.012889591977000237, -0.022219551727175713, 0.04524765908718109, 0.11818904429674149, 0.02017223834991455, 0.02498805522918701, -0.05276084691286087, -0.01206181850284338, 0.013353627175092697, -0.03493000194430351, -0.01095697283744812, -0.15088242292404175, 0.04418520629405975, 0.11013902723789215, 0.07711341977119446, -0.03398584574460983, 0.001250821864232421, -0.12282838672399521, 0.18648585677146912, -0.16439808905124664, -0.05865967646241188, -0.09643249958753586, -0.11241123080253601, -0.01130988635122776, -0.03481115400791168, 0.0324011966586113, -0.02719680406153202, 0.03895329684019089, -0.07972115278244019, -0.1737813949584961, 0.12684212625026703, -0.09612250328063965, -0.061640407890081406, -0.03569447249174118, 0.1905517876148224, -0.09809403866529465, -0.01799689419567585, 0.037167031317949295, 0.003982815425843, -0.0542159341275692, -0.11440543830394745, 0.02855151705443859, 0.04109293222427368, 0.033479753881692886, 0.0016771432710811496, -0.11700403690338135, -0.06692426651716232, -0.004046223126351833, -0.08328445255756378, 0.24640822410583496, 0.29283392429351807, -0.04596100375056267, 0.1588197648525238, 0.13218165934085846, -0.11227844655513763, -0.3360916078090668, -0.11419223248958588, -0.1741398572921753, -0.0473504364490509, -0.002386464737355709, -0.058264706283807755, 0.07159250974655151, 0.03051902912557125, -0.048579633235931396, 0.14763087034225464, -0.20566119253635406, -0.11314748972654343, 0.1453769952058792, 0.03906257823109627, 0.36354249715805054, -0.182374507188797, -0.10689529776573181, -0.12887349724769592, -0.16553707420825958, 0.13653558492660522, -0.1316424459218979, 0.094969742000103, -0.007315364200621843, 0.055984001606702805, 0.026648670434951782, -0.041909828782081604, 0.09792201220989227, -0.031702201813459396, 0.07547570765018463, -0.13294239342212677, -0.042379867285490036, 0.032361630350351334, -0.027424655854701996, 0.06689286231994629, -0.19284602999687195, 0.022383198142051697, -0.06376155465841293, -0.030110254883766174, -0.029095415025949478, 0.06940870732069016, -0.008284754119813442, -0.05276322364807129, -0.03524418920278549, -0.07097437232732773, 0.036924704909324646, -0.008823559619486332, 0.258166640996933, -0.08200068771839142, 0.1435610055923462, 0.22969910502433777, 0.16402040421962738, -0.13909795880317688, 0.07751703262329102, -0.03656279295682907, -0.0798199400305748, 0.05720395967364311, -0.15529334545135498, 0.0649019405245781, 0.09394524991512299, -0.020225821062922478, 0.09162230044603348, 0.09798222035169601, 0.00979637447744608, -0.005413372535258532, 0.1492045670747757, -0.20514246821403503, -0.1003606840968132, -0.045768361538648605, 0.08841144293546677, 0.07456500083208084, 0.10927282273769379, 0.19001200795173645, -0.02666502818465233, 0.024934252724051476, -0.0014060529647395015, 0.057744503021240234, -0.05252347141504288, 0.04748021438717842, 0.020309077575802803, 0.03647780790925026, -0.1200990304350853, 0.10943876206874847, -0.032060734927654266, -0.1143689900636673, 0.035953227430582047, 0.11986535787582397, -0.10597530007362366, -0.12283480167388916, -0.04841301962733269, 0.1376807689666748, -0.15514381229877472, -0.07959575951099396, -0.07887280732393265, -0.18433836102485657, 0.033372242003679276, 0.23955963551998138, 0.036081865429878235, 0.07665778696537018, 0.034501440823078156, -0.03135858476161957, -0.0332990400493145, 0.0450265035033226, -0.08376584202051163, 0.02836722880601883, -0.1292603313922882, 0.013662016950547695, -0.029554761946201324, 0.06653987616300583, -0.0747770220041275, -0.005165080074220896, -0.15521204471588135, 0.024890676140785217, -0.14799179136753082, 0.006287814117968082, -0.06621571630239487, -0.02280518412590027, 0.0076015302911400795, -0.012781600467860699, -0.056757766753435135, -0.020954884588718414, -0.10184937715530396, 0.011013524606823921, -0.02370467223227024, 0.048052769154310226, -0.08744794875383377, -0.03087439015507698, 0.041023094207048416, -0.04014357924461365, 0.10149496048688889, 0.0686996653676033, -0.08736540377140045, 0.11272776126861572, -0.22351643443107605, -0.036217331886291504, 0.13396640121936798, 0.0044109211303293705, 0.02989373169839382, 0.05103369057178497, -0.03830115497112274, 0.10754726082086563, 0.017391212284564972, 0.03653336688876152, -0.002329655224457383, -0.10563276708126068, 0.024246202781796455, -0.04161437973380089, -0.10984683781862259, -0.018942339345812798, -0.06254519522190094, 0.06314903497695923, -0.03017268143594265, 0.14784294366836548, -0.08171480149030685, 0.037169553339481354, -0.0517251193523407, 0.03659065440297127, 0.036367785185575485, -0.16036273539066315, -0.08944962918758392, -0.06599938869476318, 0.007963553071022034, -0.03353995084762573, 0.25739243626594543, -0.002668095054104924, -0.07485136389732361, 0.07247384637594223, 0.01097193080931902, 0.013616526499390602, 0.02459101565182209, 0.2530827522277832, 0.03458145633339882, -0.01872875913977623, -0.1476380079984665, 0.015492787584662437, 0.04218541830778122, -0.06074988469481468, 0.11544693261384964, 0.0993463471531868, -0.07081745564937592, 0.11420585215091705, 0.08185305446386337, -0.027696184813976288, -0.046162061393260956, -0.08060360699892044, -0.03348478674888611, 0.06855632364749908, -0.03132452070713043, 0.07537361234426498, 0.22101734578609467, -0.01775355450809002, -0.0011685204226523638, -0.06780751049518585, -0.031885482370853424, -0.1863650232553482, -0.0841546580195427, -0.10429336130619049, -0.13841724395751953, 0.02425653301179409, -0.09412970393896103, 0.07430317252874374, 0.019939376041293144, 0.05734672397375107, -0.03683450445532799, 0.0940014198422432, 0.0004930990980938077, -0.05851813405752182, 0.01835847832262516, -0.026307005435228348, 0.025924304500222206, -0.03434149920940399, -0.021814318373799324, -0.042366135865449905, -0.07287942618131638, -0.04727296903729439, 0.07000782340765, 0.010771182365715504, 0.03856037184596062, -0.12262279540300369, -0.05158356577157974, -0.025341777130961418, 0.08332459628582001, 0.00042757162009365857, 0.1390371024608612, 0.019977936521172523, -0.038895536214113235, 0.06657611578702927, 0.19378311932086945, -0.08377885073423386, -0.114676833152771, 0.010468890890479088, 0.18122731149196625, 0.05526398867368698, 0.138681560754776, -0.049045998603105545, -0.013016736134886742, -0.008810778148472309, 0.2937930226325989, 0.257145494222641, -0.06495130807161331, 0.04491841420531273, -0.062360212206840515, 0.04619316756725311, 0.05555616319179535, 0.07588943094015121, 0.10346000641584396, 0.2828470766544342, -0.04612942412495613, -0.02748578041791916, -0.007263211067765951, 0.01569332368671894, -0.1111651286482811, 0.09446040540933609, -0.0144914910197258, -0.054909370839595795, -0.01669856160879135, 0.06318347901105881, -0.15636524558067322, 0.056414615362882614, -0.06769883632659912, -0.15390802919864655, -0.008484174497425556, -0.0011523261200636625, 0.15808144211769104, -0.006652408745139837, 0.05238869786262512, -0.03263389319181442, -0.09687774628400803, 0.02537815272808075, 0.006523885764181614, -0.2290477752685547, 0.027606386691331863, 0.012579442001879215, -0.05905579775571823, 0.05620573088526726, -0.00009738940570969135, 0.0543033704161644, 0.07550898939371109, 0.029322348535060883, -0.044229455292224884, 0.1737268567085266, 0.027487516403198242, -0.11136257648468018, 0.041999392211437225, -0.004863246344029903, -0.008640053682029247, 0.019216388463974, 0.04285409674048424, -0.09746899455785751, 0.05180269479751587, 0.07809833437204361, -0.11871751397848129, -0.023186665028333664, 0.024502970278263092, -0.070523202419281, 0.0801381766796112, 0.03692895546555519, -0.026914743706583977, 0.004324541427195072, -0.021714936941862106, 0.023277346044778824, -0.020302291959524155, -0.12371190637350082, -0.009685642085969448, -0.13206802308559418, -0.03183959424495697, 0.1363690197467804, 0.016948679462075233, -0.27292850613594055, 0.003331792773678899, -0.08759225904941559, 0.05062190443277359, -0.17548513412475586, 0.04686470702290535, 0.20208954811096191, 0.012469199486076832, -0.02944374457001686, -0.175132617354393, 0.05326329916715622, 0.09063606709241867, -0.06020814925432205, -0.11170734465122223 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # roberta-base-detect-cheapfake-combined-train-test-context This model is a fine-tuned version of [roberta-base](https://huggingface.co/roberta-base) on the None dataset. It achieves the following results on the evaluation set: - Loss: 0.4470 - Accuracy: 0.78 - F1: 0.7442 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 5e-06 - train_batch_size: 32 - eval_batch_size: 32 - seed: 42 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - num_epochs: 5 ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | F1 | |:-------------:|:-----:|:----:|:---------------:|:--------:|:------:| | No log | 1.0 | 75 | 0.6170 | 0.49 | 0.6577 | | No log | 2.0 | 150 | 0.4459 | 0.79 | 0.7692 | | No log | 3.0 | 225 | 0.4441 | 0.79 | 0.7692 | | No log | 4.0 | 300 | 0.4404 | 0.81 | 0.7865 | | No log | 5.0 | 375 | 0.4470 | 0.78 | 0.7442 | ### Framework versions - Transformers 4.37.0 - Pytorch 2.1.2 - Datasets 2.1.0 - Tokenizers 0.15.1
{"license": "mit", "tags": ["generated_from_trainer"], "metrics": ["accuracy", "f1"], "base_model": "roberta-base", "model-index": [{"name": "roberta-base-detect-cheapfake-combined-train-test-context", "results": []}]}
text-classification
hoanghoavienvo/roberta-base-detect-cheapfake-combined-train-test-context
[ "transformers", "tensorboard", "safetensors", "roberta", "text-classification", "generated_from_trainer", "base_model:roberta-base", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
2024-02-10T16:21:31+00:00
[]
[]
TAGS #transformers #tensorboard #safetensors #roberta #text-classification #generated_from_trainer #base_model-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us
roberta-base-detect-cheapfake-combined-train-test-context ========================================================= This model is a fine-tuned version of roberta-base on the None dataset. It achieves the following results on the evaluation set: * Loss: 0.4470 * Accuracy: 0.78 * F1: 0.7442 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 5e-06 * train\_batch\_size: 32 * eval\_batch\_size: 32 * seed: 42 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * num\_epochs: 5 ### Training results ### Framework versions * Transformers 4.37.0 * Pytorch 2.1.2 * Datasets 2.1.0 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.1.0\n* Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #tensorboard #safetensors #roberta #text-classification #generated_from_trainer #base_model-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.1.0\n* Tokenizers 0.15.1" ]
[ 63, 98, 4, 30 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #roberta #text-classification #generated_from_trainer #base_model-roberta-base #license-mit #autotrain_compatible #endpoints_compatible #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 5e-06\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 32\n* seed: 42\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* num\\_epochs: 5### Training results### Framework versions\n\n\n* Transformers 4.37.0\n* Pytorch 2.1.2\n* Datasets 2.1.0\n* Tokenizers 0.15.1" ]
[ -0.08621460944414139, 0.07034635543823242, -0.0019199398811906576, 0.10115745663642883, 0.16650773584842682, 0.014715912751853466, 0.16630806028842926, 0.10816951096057892, -0.10288658738136292, 0.0370427630841732, 0.12479393929243088, 0.15665750205516815, -0.00386789720505476, 0.13193999230861664, -0.07780896127223969, -0.2460506111383438, 0.0017037739744409919, 0.03195207566022873, -0.08352852612733841, 0.11412577331066132, 0.10524716228246689, -0.13441744446754456, 0.08543899655342102, -0.010788877494633198, -0.20579494535923004, 0.03757031634449959, 0.04433418810367584, -0.06682558357715607, 0.13942250609397888, 0.04353126883506775, 0.138076514005661, 0.03182223439216614, 0.08476164191961288, -0.19053460657596588, 0.019515322521328926, 0.06025005504488945, -0.01696426421403885, 0.08433520048856735, 0.04447320103645325, -0.02846045419573784, 0.11073671281337738, -0.09475521743297577, 0.06298670917749405, 0.02233942784368992, -0.13000643253326416, -0.20548291504383087, -0.07241541147232056, 0.03683169558644295, 0.09002912789583206, 0.08057185262441635, -0.018453553318977356, 0.15328148007392883, -0.060680683702230453, 0.10176348686218262, 0.2039148211479187, -0.3121492266654968, -0.06807814538478851, 0.0557270273566246, 0.026285406202077866, 0.08735282719135284, -0.115474633872509, 0.0026458969805389643, 0.07811914384365082, 0.020640091970562935, 0.12106876820325851, -0.037907227873802185, -0.05945747345685959, 0.01250428706407547, -0.1404155045747757, 0.004329723306000233, 0.14547623693943024, 0.04535907506942749, -0.04377753287553787, -0.04129227623343468, -0.05750489607453346, -0.12297330051660538, -0.041320838034152985, -0.03006129153072834, 0.04189131036400795, -0.028421731665730476, -0.11054307967424393, -0.017965661361813545, -0.1175297349691391, -0.07437912374734879, -0.06167883425951004, 0.1773151457309723, 0.03495987877249718, 0.0025804294273257256, -0.03309919312596321, 0.09120360016822815, -0.038548119366168976, -0.12727007269859314, 0.01885819435119629, 0.023834383115172386, 0.004928858485072851, -0.0724116712808609, -0.06592026352882385, -0.09593129903078079, 0.03041834942996502, 0.15105116367340088, -0.06584733724594116, 0.050920501351356506, 0.00844074971973896, 0.043425071984529495, -0.09286519140005112, 0.156424880027771, -0.04224889352917671, -0.0200910996645689, 0.02856389805674553, 0.05661073699593544, 0.039455950260162354, -0.002625084714964032, -0.12428244203329086, 0.018618585541844368, 0.11993711441755295, 0.014321994967758656, -0.0776013731956482, 0.08416330814361572, -0.03710022568702698, 0.004922328982502222, 0.012066755443811417, -0.09140391647815704, 0.03385770693421364, 0.003724277252331376, -0.060643598437309265, -0.06862009316682816, 0.025731030851602554, 0.01758686453104019, 0.010992304421961308, 0.10249429941177368, -0.09221751987934113, 0.02013399265706539, -0.09168679267168045, -0.13208584487438202, 0.009481038898229599, -0.06193193048238754, 0.03720010817050934, -0.11700829863548279, -0.15841850638389587, -0.01614351198077202, 0.04650472104549408, -0.026190951466560364, -0.01897256076335907, -0.054438266903162, -0.08079638332128525, 0.012490568682551384, -0.015759531408548355, 0.09302414953708649, -0.055287234485149384, 0.09891168028116226, 0.06301828473806381, 0.0675201267004013, -0.05937451496720314, 0.031105026602745056, -0.10197152197360992, 0.014138037338852882, -0.20764753222465515, 0.02310292422771454, -0.06100793182849884, 0.07491709291934967, -0.07645316421985626, -0.08150498569011688, -0.005314336623996496, 0.024412302300333977, 0.07218009233474731, 0.08865291625261307, -0.14711210131645203, -0.07527957856655121, 0.16828382015228271, -0.10392335057258606, -0.1358744204044342, 0.11798913776874542, -0.0721166729927063, 0.07070565968751907, 0.0756799504160881, 0.19184152781963348, 0.0618230402469635, -0.0861574038863182, 0.005990234669297934, -0.012431567534804344, 0.0401887446641922, -0.05041053146123886, 0.05824347212910652, 0.008431577123701572, -0.0034512151032686234, 0.011282360181212425, -0.02325744368135929, 0.048965733498334885, -0.09330045431852341, -0.07874763011932373, -0.03844728320837021, -0.10583189874887466, 0.05032626539468765, 0.06182317063212395, 0.08372177183628082, -0.13273939490318298, -0.08757881075143814, 0.10252247005701065, 0.0725402757525444, -0.07010117918252945, 0.018913112580776215, -0.07962192595005035, 0.06808958947658539, -0.05985528603196144, -0.029801342636346817, -0.16415730118751526, -0.05009562149643898, -0.002676796168088913, 0.031824056059122086, 0.0347660593688488, 0.014260555617511272, 0.08079540729522705, 0.07340314984321594, -0.0760706290602684, -0.02627664804458618, 0.001719681778922677, 0.015024897642433643, -0.1262606680393219, -0.20914557576179504, -0.0023294438142329454, -0.04222992807626724, 0.1321078985929489, -0.25167644023895264, 0.05542728677392006, 0.0058716339990496635, 0.08717125654220581, 0.04098164662718773, 0.002394840121269226, -0.04215797781944275, 0.07370348274707794, -0.05050470307469368, -0.05236615985631943, 0.05169191583991051, 0.005766124464571476, -0.09053945541381836, -0.04152698442339897, -0.15880395472049713, 0.19915197789669037, 0.14128848910331726, -0.11368162930011749, -0.10693640261888504, -0.0018688521813601255, -0.03975243493914604, -0.021979743614792824, -0.048833996057510376, 0.015094748698174953, 0.12209979444742203, -0.0269149336963892, 0.15366685390472412, -0.07238663733005524, -0.03532639890909195, 0.020037991926074028, -0.06456677615642548, 0.009610732086002827, 0.10727270692586899, 0.08742791414260864, -0.13918396830558777, 0.15090174973011017, 0.14885742962360382, -0.10853871703147888, 0.1527588665485382, -0.0332222580909729, -0.05289981886744499, -0.02175251953303814, 0.0012371899792924523, 0.015059403143823147, 0.109502412378788, -0.11255775392055511, -0.01091880351305008, -0.0013665318256244063, 0.007489899173378944, 0.021909918636083603, -0.2227829545736313, -0.03621427342295647, 0.029583213850855827, -0.03281821683049202, 0.013786832801997662, -0.019375290721654892, -0.00892646238207817, 0.10360633581876755, -0.004862393252551556, -0.08221158385276794, 0.04331734776496887, 0.0006397544057108462, -0.09185978025197983, 0.22285795211791992, -0.07136739790439606, -0.11943015456199646, -0.13255707919597626, -0.06425787508487701, -0.04347159340977669, 0.033622320741415024, 0.06763775646686554, -0.08547840267419815, -0.03778986632823944, -0.09566111862659454, 0.007720544934272766, 0.02329927682876587, 0.0344257727265358, -0.0016292420914396644, 0.015377731062471867, 0.07979975640773773, -0.10849437862634659, -0.008027816191315651, -0.05612938851118088, -0.07000115513801575, 0.045762114226818085, 0.030557934194803238, 0.11570652574300766, 0.1440325677394867, -0.04411771893501282, -0.0012083378387615085, -0.043645355850458145, 0.22117897868156433, -0.06906338036060333, -0.017804423347115517, 0.12098298966884613, -0.018330426886677742, 0.040507640689611435, 0.1338038295507431, 0.06436765938997269, -0.09637270122766495, 0.03533366322517395, 0.04467233270406723, -0.032245371490716934, -0.2173183113336563, -0.029396653175354004, -0.028981659561395645, -0.012439590878784657, 0.0846438854932785, 0.041100725531578064, 0.048482928425073624, 0.0778949186205864, 0.03482109308242798, 0.06626007705926895, -0.006496085785329342, 0.07564367353916168, 0.10486944764852524, 0.04796665534377098, 0.13633906841278076, -0.05979489907622337, -0.07824981957674026, 0.027423327788710594, -0.016282638534903526, 0.19207338988780975, 0.02345864474773407, 0.10902959853410721, 0.05600495636463165, 0.14305923879146576, 0.012536057271063328, 0.06671948730945587, -0.0012640421045944095, -0.06476438790559769, -0.0017678000731393695, -0.04603888466954231, -0.011062346398830414, 0.0424518920481205, -0.09809726476669312, 0.05192866176366806, -0.11623319983482361, 0.014843374490737915, 0.06703438609838486, 0.20600728690624237, 0.057366516441106796, -0.32683876156806946, -0.0957888662815094, 0.027665939182043076, -0.02500305138528347, -0.016003021970391273, 0.02404729276895523, 0.11640861630439758, -0.04653707519173622, 0.04369260370731354, -0.06907264143228531, 0.0758499950170517, -0.025330089032649994, 0.04277022182941437, 0.03960750624537468, 0.09704454243183136, -0.02770659327507019, 0.06778678297996521, -0.2931763529777527, 0.2799687683582306, 0.02005266211926937, 0.08785542100667953, -0.0463939793407917, -0.00894091371446848, 0.030869992449879646, 0.10063181072473526, 0.06270267069339752, -0.030952533707022667, -0.10102472454309464, -0.19663341343402863, -0.030363373458385468, 0.03185392543673515, 0.1124514639377594, -0.017430542036890984, 0.10546272993087769, -0.03484777361154556, 0.0047864909283816814, 0.09675119072198868, -0.03662927448749542, -0.09195798635482788, -0.08765549212694168, -0.03306208923459053, 0.027491334825754166, -0.029358653351664543, -0.08068525046110153, -0.10269691795110703, -0.1264357715845108, 0.17448873817920685, -0.028650738298892975, -0.015903662890195847, -0.1050434410572052, 0.08985832333564758, 0.03873691335320473, -0.08446221053600311, 0.041750937700271606, 0.015750380232930183, 0.08344025909900665, 0.021724918857216835, -0.06324604898691177, 0.1342329978942871, -0.06491038203239441, -0.16354233026504517, -0.062297314405441284, 0.09393078833818436, 0.0215628519654274, 0.041513506323099136, 0.006853341590613127, 0.012888304889202118, -0.009470623917877674, -0.07694359123706818, 0.030128495767712593, -0.036660972982645035, 0.05633040890097618, 0.016162028536200523, -0.06458757817745209, -0.02503466233611107, -0.06137208268046379, -0.031021272763609886, 0.16852742433547974, 0.2912300229072571, -0.08650317043066025, -0.014204743318259716, 0.06477013975381851, -0.0666562020778656, -0.21708612143993378, 0.0695045068860054, 0.015195329673588276, 0.00017304415814578533, 0.05297519639134407, -0.14020195603370667, 0.12152352929115295, 0.09696492552757263, -0.02311311848461628, 0.09208936989307404, -0.2660888433456421, -0.14487594366073608, 0.13833503425121307, 0.16948775947093964, 0.13819974660873413, -0.16753290593624115, -0.028249239549040794, -0.041832804679870605, -0.07565026730298996, 0.10307412594556808, -0.15185464918613434, 0.11157113313674927, 0.0009330498287454247, 0.05784858763217926, 0.005438691005110741, -0.05826614424586296, 0.11652719229459763, -0.017209693789482117, 0.12986524403095245, -0.06930266320705414, -0.03487389534711838, 0.06814896315336227, -0.047473613172769547, 0.018708450719714165, -0.09876470267772675, 0.03085518628358841, -0.04120754078030586, -0.03367144614458084, -0.054560791701078415, 0.05033528432250023, -0.036152809858322144, -0.07031293213367462, -0.04981227219104767, 0.03086826577782631, 0.019111763685941696, -0.01952076144516468, 0.17663782835006714, -0.0020540719851851463, 0.1839490383863449, 0.14268068969249725, 0.09166259318590164, -0.057264018803834915, 0.011308937333524227, 0.00958600640296936, -0.042966630309820175, 0.05829288437962532, -0.15859676897525787, 0.03947537764906883, 0.11097300052642822, 0.011818347498774529, 0.15359733998775482, 0.08246209472417831, -0.03407113626599312, 0.029384920373558998, 0.08883360028266907, -0.16351072490215302, -0.11217951029539108, -0.00722486712038517, -0.0611744225025177, -0.10050814598798752, 0.08688836544752121, 0.12312385439872742, -0.07501193135976791, 0.0005258176242932677, -0.01341442484408617, 0.00048384827096015215, -0.0507504940032959, 0.18453915417194366, 0.09220081567764282, 0.045537300407886505, -0.0735415443778038, 0.07366827875375748, 0.041846778243780136, -0.06537298858165741, 0.01792643405497074, 0.05824901536107063, -0.07235166430473328, -0.0520692877471447, 0.07197844237089157, 0.21372637152671814, -0.06219947338104248, -0.05331191048026085, -0.1639651507139206, -0.11016421020030975, 0.0420074425637722, 0.21725662052631378, 0.10022244602441788, 0.0015682053053751588, -0.02477157488465309, 0.02682790905237198, -0.13826236128807068, 0.1056167259812355, 0.035438958555459976, 0.08831042796373367, -0.15440605580806732, 0.16504178941249847, -0.004934485536068678, 0.00738111324608326, -0.03403133153915405, 0.03900950029492378, -0.13402023911476135, -0.001358143868856132, -0.12975062429904938, -0.017948977649211884, -0.036150187253952026, 0.00705350935459137, 0.011725328862667084, -0.05977802351117134, -0.073634073138237, 0.011922621168196201, -0.10703765600919724, -0.006815774366259575, 0.0448417030274868, 0.04978194087743759, -0.13038961589336395, -0.035286642611026764, 0.013649185188114643, -0.05796428769826889, 0.0619937889277935, 0.01848706416785717, 0.021011264994740486, 0.06734245270490646, -0.19985172152519226, 0.03395448997616768, 0.07704784721136093, -0.012239503674209118, 0.050021715462207794, -0.07495354861021042, -0.005170348100364208, -0.000559842272195965, 0.0739847794175148, 0.024887854233384132, 0.0753970667719841, -0.12177865952253342, 0.021222397685050964, -0.04275210574269295, -0.07630109041929245, -0.061168380081653595, 0.035573262721300125, 0.08181490004062653, -0.009855812415480614, 0.2004423588514328, -0.11339035630226135, 0.005301987286657095, -0.20558036863803864, 0.01143474131822586, -0.0068168132565915585, -0.11917541176080704, -0.11895749717950821, -0.05339057371020317, 0.0550699420273304, -0.06255640834569931, 0.14445756375789642, 0.027589716017246246, 0.022842060774564743, 0.04499496892094612, -0.04698629304766655, 0.04499046131968498, 0.03900059312582016, 0.23826658725738525, 0.022651933133602142, -0.040031518787145615, 0.011317627504467964, 0.05503624305129051, 0.1177772507071495, 0.07435490936040878, 0.18096236884593964, 0.16575883328914642, -0.07952086627483368, 0.10649328678846359, 0.04639053717255592, -0.04334588348865509, -0.11245573312044144, 0.05445774644613266, -0.039720553904771805, 0.06882435083389282, -0.028617139905691147, 0.16729699075222015, 0.13477405905723572, -0.15014676749706268, 0.008578702807426453, -0.06372388452291489, -0.08196338266134262, -0.1247776448726654, -0.045366864651441574, -0.11185631901025772, -0.15185493230819702, 0.00907887052744627, -0.11548378318548203, 0.0030371355824172497, 0.08570306748151779, 0.009196741506457329, -0.022737177088856697, 0.1770169734954834, 0.017320849001407623, 0.03781522065401077, 0.04434143751859665, -0.005049495026469231, -0.034411150962114334, -0.08074658364057541, -0.07970217615365982, 0.0009836758254095912, -0.03258924558758736, 0.011286775581538677, -0.05255812406539917, -0.03520456701517105, 0.04501992464065552, -0.02387145720422268, -0.10394816100597382, 0.016458895057439804, 0.0423688068985939, 0.05770774930715561, 0.03840900957584381, 0.012109171599149704, -0.003801998682320118, -0.0024740351364016533, 0.23563319444656372, -0.07412221282720566, -0.07752007246017456, -0.09853071719408035, 0.2599938213825226, 0.0445479080080986, 0.031177973374724388, 0.0011631363304331899, -0.1040104404091835, 0.038260821253061295, 0.21774843335151672, 0.20907457172870636, -0.07038480788469315, 0.01571628637611866, -0.020128199830651283, -0.010362588800489902, -0.03143841028213501, 0.0996226817369461, 0.10291974991559982, 0.018538620322942734, -0.08051638305187225, -0.051019784063100815, -0.03268466889858246, -0.002526161726564169, -0.04672601819038391, 0.06364256143569946, 0.046502791345119476, 0.019397519528865814, -0.048147737979888916, 0.06624621897935867, -0.03716207295656204, -0.11400065571069717, 0.056192513555288315, -0.193911612033844, -0.14073801040649414, -0.011184628121554852, 0.11281964927911758, -0.01577158086001873, 0.06394682079553604, -0.03357775881886482, -0.004784015938639641, 0.02422148361802101, -0.016762642189860344, -0.07807765901088715, -0.06770554929971695, 0.06420914828777313, -0.10638224333524704, 0.2148200273513794, -0.04502749443054199, 0.042727094143629074, 0.1441943198442459, 0.034243207424879074, -0.06757210195064545, 0.10389722883701324, 0.03781982138752937, -0.07656672596931458, 0.042372412979602814, 0.09151187539100647, -0.04608732461929321, 0.10941384732723236, 0.05434328690171242, -0.14807142317295074, 0.029908789321780205, -0.09365039318799973, -0.08575345575809479, -0.0612005740404129, -0.03713468089699745, -0.06510815024375916, 0.12557339668273926, 0.18704690039157867, -0.029371334239840508, 0.0244159996509552, -0.05118292570114136, 0.02412959188222885, 0.07948243618011475, 0.027997558936476707, -0.044953200966119766, -0.25250717997550964, 0.02982354536652565, 0.09052589535713196, -0.015610013157129288, -0.3069535791873932, -0.07393258064985275, -0.011071725748479366, -0.04628496244549751, -0.09845677763223648, 0.08551959693431854, 0.1397314816713333, 0.05784335732460022, -0.06320219486951828, -0.12059269845485687, -0.07806798070669174, 0.16363604366779327, -0.1351093053817749, -0.10535386204719543 ]
null
null
transformers
<!-- This model card has been generated automatically according to the information the Trainer had access to. You should probably proofread and complete it, then remove this comment. --> # smolm-autoreg-bpe-counterfactual-babylm-only_measure_nps_as_singular_removal-seed_1024-1e-3 This model was trained from scratch on the kanishka/counterfactual-babylm-only_measure_nps_as_singular_removal dataset. It achieves the following results on the evaluation set: - Loss: 3.4259 - Accuracy: 0.4097 ## Model description More information needed ## Intended uses & limitations More information needed ## Training and evaluation data More information needed ## Training procedure ### Training hyperparameters The following hyperparameters were used during training: - learning_rate: 0.001 - train_batch_size: 32 - eval_batch_size: 64 - seed: 1024 - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 - lr_scheduler_type: linear - lr_scheduler_warmup_steps: 32000 - num_epochs: 20.0 - mixed_precision_training: Native AMP ### Training results | Training Loss | Epoch | Step | Validation Loss | Accuracy | |:-------------:|:-----:|:------:|:---------------:|:--------:| | 3.6017 | 1.0 | 18600 | 3.7683 | 0.3593 | | 3.3799 | 2.0 | 37200 | 3.5935 | 0.3790 | | 3.2546 | 3.0 | 55800 | 3.4823 | 0.3915 | | 3.1737 | 4.0 | 74400 | 3.4548 | 0.3978 | | 3.1178 | 5.0 | 93000 | 3.4163 | 0.4014 | | 3.0736 | 6.0 | 111600 | 3.4017 | 0.4038 | | 3.0385 | 7.0 | 130200 | 3.3798 | 0.4057 | | 3.0068 | 8.0 | 148800 | 3.3988 | 0.4060 | | 2.9774 | 9.0 | 167400 | 3.3728 | 0.4074 | | 2.9558 | 10.0 | 186000 | 3.3695 | 0.4087 | | 2.9289 | 11.0 | 204600 | 3.3649 | 0.4094 | | 2.9058 | 12.0 | 223200 | 3.3604 | 0.4095 | | 2.8805 | 13.0 | 241800 | 3.3801 | 0.4098 | | 2.8621 | 14.0 | 260400 | 3.3871 | 0.4095 | | 2.8423 | 15.0 | 279000 | 3.3872 | 0.4096 | | 2.8216 | 16.0 | 297600 | 3.3996 | 0.4097 | | 2.8042 | 17.0 | 316200 | 3.3987 | 0.4101 | | 2.7834 | 18.0 | 334800 | 3.4020 | 0.4101 | | 2.7643 | 19.0 | 353400 | 3.4199 | 0.4097 | | 2.7463 | 20.0 | 372000 | 3.4259 | 0.4097 | ### Framework versions - Transformers 4.37.2 - Pytorch 2.1.0+cu121 - Datasets 2.16.1 - Tokenizers 0.15.1
{"tags": ["generated_from_trainer"], "datasets": ["kanishka/counterfactual-babylm-only_measure_nps_as_singular_removal"], "metrics": ["accuracy"], "model-index": [{"name": "smolm-autoreg-bpe-counterfactual-babylm-only_measure_nps_as_singular_removal-seed_1024-1e-3", "results": [{"task": {"type": "text-generation", "name": "Causal Language Modeling"}, "dataset": {"name": "kanishka/counterfactual-babylm-only_measure_nps_as_singular_removal", "type": "kanishka/counterfactual-babylm-only_measure_nps_as_singular_removal"}, "metrics": [{"type": "accuracy", "value": 0.4096600918317765, "name": "Accuracy"}]}]}]}
text-generation
kanishka/smolm-autoreg-bpe-counterfactual-babylm-only_measure_nps_as_singular_removal-seed_1024-1e-3
[ "transformers", "tensorboard", "safetensors", "opt", "text-generation", "generated_from_trainer", "dataset:kanishka/counterfactual-babylm-only_measure_nps_as_singular_removal", "model-index", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T16:25:21+00:00
[]
[]
TAGS #transformers #tensorboard #safetensors #opt #text-generation #generated_from_trainer #dataset-kanishka/counterfactual-babylm-only_measure_nps_as_singular_removal #model-index #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
smolm-autoreg-bpe-counterfactual-babylm-only\_measure\_nps\_as\_singular\_removal-seed\_1024-1e-3 ================================================================================================= This model was trained from scratch on the kanishka/counterfactual-babylm-only\_measure\_nps\_as\_singular\_removal dataset. It achieves the following results on the evaluation set: * Loss: 3.4259 * Accuracy: 0.4097 Model description ----------------- More information needed Intended uses & limitations --------------------------- More information needed Training and evaluation data ---------------------------- More information needed Training procedure ------------------ ### Training hyperparameters The following hyperparameters were used during training: * learning\_rate: 0.001 * train\_batch\_size: 32 * eval\_batch\_size: 64 * seed: 1024 * optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 * lr\_scheduler\_type: linear * lr\_scheduler\_warmup\_steps: 32000 * num\_epochs: 20.0 * mixed\_precision\_training: Native AMP ### Training results ### Framework versions * Transformers 4.37.2 * Pytorch 2.1.0+cu121 * Datasets 2.16.1 * Tokenizers 0.15.1
[ "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 64\n* seed: 1024\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 32000\n* num\\_epochs: 20.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1" ]
[ "TAGS\n#transformers #tensorboard #safetensors #opt #text-generation #generated_from_trainer #dataset-kanishka/counterfactual-babylm-only_measure_nps_as_singular_removal #model-index #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 64\n* seed: 1024\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 32000\n* num\\_epochs: 20.0\n* mixed\\_precision\\_training: Native AMP", "### Training results", "### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1" ]
[ 93, 132, 4, 33 ]
[ "passage: TAGS\n#transformers #tensorboard #safetensors #opt #text-generation #generated_from_trainer #dataset-kanishka/counterfactual-babylm-only_measure_nps_as_singular_removal #model-index #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n### Training hyperparameters\n\n\nThe following hyperparameters were used during training:\n\n\n* learning\\_rate: 0.001\n* train\\_batch\\_size: 32\n* eval\\_batch\\_size: 64\n* seed: 1024\n* optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08\n* lr\\_scheduler\\_type: linear\n* lr\\_scheduler\\_warmup\\_steps: 32000\n* num\\_epochs: 20.0\n* mixed\\_precision\\_training: Native AMP### Training results### Framework versions\n\n\n* Transformers 4.37.2\n* Pytorch 2.1.0+cu121\n* Datasets 2.16.1\n* Tokenizers 0.15.1" ]
[ -0.11735479533672333, 0.10444210469722748, -0.0031074327416718006, 0.056101709604263306, 0.11266118288040161, 0.02413639985024929, 0.1589384526014328, 0.14130449295043945, -0.0636143758893013, 0.09187088906764984, 0.13230431079864502, 0.07454783469438553, 0.05784912407398224, 0.1282913088798523, -0.05120622366666794, -0.2699681222438812, 0.025921914726495743, 0.02278462052345276, -0.07315012067556381, 0.12000881135463715, 0.08911918103694916, -0.11337532103061676, 0.05321149900555611, 0.01822592131793499, -0.11960698664188385, -0.012931362725794315, -0.0055831726640462875, -0.06171180307865143, 0.10264512151479721, 0.01502537727355957, 0.1241513192653656, 0.028840605169534683, 0.08054423332214355, -0.2181701809167862, 0.013859782367944717, 0.06610953062772751, 0.02833361178636551, 0.09025254100561142, 0.09394490718841553, -0.021046077832579613, 0.1060667484998703, -0.10147811472415924, 0.07912284880876541, 0.03925229609012604, -0.11740629374980927, -0.24627043306827545, -0.06711205840110779, 0.0551719032227993, 0.10386255383491516, 0.07800383120775223, -0.02275681495666504, 0.10952626168727875, -0.04113166406750679, 0.09269563853740692, 0.19469726085662842, -0.22349177300930023, -0.09505727142095566, -0.02942012809216976, 0.055131979286670685, 0.07059089094400406, -0.1146169975399971, -0.02249796874821186, 0.04120386391878128, 0.020737631246447563, 0.11590875685214996, 0.007635165471583605, 0.050133489072322845, -0.02001442015171051, -0.1411886215209961, -0.0627627819776535, 0.13182564079761505, 0.06939318031072617, -0.04268042743206024, -0.09374381601810455, -0.04747818410396576, -0.17137287557125092, -0.055926840752363205, 0.018807020038366318, 0.01690099574625492, -0.039739664644002914, -0.08993200957775116, -0.022701406851410866, -0.09834794700145721, -0.08541475981473923, 0.029250049963593483, 0.18555627763271332, 0.05546918138861656, -0.030943436548113823, -0.01786608248949051, 0.10984303802251816, 0.07148953527212143, -0.14672881364822388, -0.002935541793704033, 0.025485927239060402, -0.06272022426128387, -0.029557082802057266, -0.03143242374062538, -0.046068958938121796, 0.014555279165506363, 0.12565475702285767, -0.03740117698907852, 0.0841824933886528, 0.008916226215660572, 0.03383684158325195, -0.078300341963768, 0.1521056592464447, -0.030816705897450447, 0.010398345068097115, -0.024683700874447823, 0.13594496250152588, 0.0018687311094254255, -0.011428321711719036, -0.035725418478250504, 0.015337507240474224, 0.14016233384609222, 0.0425579808652401, -0.014046378433704376, 0.04259074851870537, -0.0645691379904747, -0.020161544904112816, 0.011238627135753632, -0.10543905943632126, 0.027528949081897736, 0.026257473975419998, -0.04377971589565277, -0.010958398692309856, 0.01671847701072693, 0.004706535954028368, -0.014311338774859905, 0.08994639664888382, -0.08560988306999207, -0.0035045400727540255, -0.08806859701871872, -0.08303385227918625, 0.00899939052760601, -0.040364041924476624, 0.005936742294579744, -0.09473586082458496, -0.13676078617572784, -0.04746796190738678, 0.04868611320853233, -0.03547482192516327, -0.045661214739084244, -0.050979260355234146, -0.07832310348749161, 0.051039934158325195, -0.013378314673900604, 0.11513012647628784, -0.053476061671972275, 0.1058826893568039, 0.03040085732936859, 0.037339773029088974, 0.04914754629135132, 0.0422980859875679, -0.06818599998950958, 0.0680161789059639, -0.09950876235961914, 0.0661059021949768, -0.07706129550933838, 0.043176133185625076, -0.11681653559207916, -0.11052201688289642, -0.03988289088010788, -0.002764672040939331, 0.0901026651263237, 0.11700291931629181, -0.14378952980041504, -0.06595705449581146, 0.1832171529531479, -0.09657593071460724, -0.1287403553724289, 0.1185515746474266, -0.029671097174286842, 0.031061014160513878, 0.03921922296285629, 0.15093699097633362, 0.08247105777263641, -0.07894290238618851, -0.00007327285129576921, -0.04533304646611214, 0.09665972739458084, 0.03317159041762352, 0.10150068253278732, -0.012399262748658657, -0.026471184566617012, -0.007530935574322939, -0.05109962821006775, 0.06705804169178009, -0.09093787521123886, -0.08309066295623779, -0.0386669896543026, -0.08154786378145218, 0.025870902463793755, 0.05897754058241844, 0.03780052065849304, -0.08757723867893219, -0.11462590843439102, 0.03248616307973862, 0.11078870296478271, -0.10574416816234589, 0.011445018462836742, -0.06764411181211472, 0.032850079238414764, -0.06646408885717392, -0.013931847177445889, -0.1484798938035965, -0.08668289333581924, 0.03405285254120827, -0.05107180401682854, -0.019075853750109673, -0.04889286309480667, 0.09152359515428543, 0.062364403158426285, -0.05249089375138283, -0.09292706102132797, -0.06547655910253525, 0.0030254684388637543, -0.09121298044919968, -0.201995387673378, -0.06985736638307571, -0.034840162843465805, 0.17191749811172485, -0.24920900166034698, 0.04014279320836067, -0.012319475412368774, 0.11914151161909103, 0.04199730232357979, -0.049587786197662354, 0.0043680802918970585, 0.04068869352340698, -0.03572125360369682, -0.09042147547006607, 0.04563277214765549, 0.013608399778604507, -0.11731986701488495, 0.031457237899303436, -0.14944660663604736, 0.08536297082901001, 0.09325333684682846, -0.008628771640360355, -0.09306448698043823, -0.04923829063773155, -0.07547475397586823, -0.06121378391981125, -0.03630401939153671, -0.01647225394845009, 0.15298008918762207, 0.03456978127360344, 0.1283898800611496, -0.09973685443401337, -0.054143212735652924, 0.02420113794505596, -0.01862644962966442, -0.02440009079873562, 0.1399523913860321, 0.04753801226615906, -0.10523388534784317, 0.09616550803184509, 0.10439255088567734, -0.07590355724096298, 0.15661057829856873, -0.059769853949546814, -0.1155627965927124, -0.02271808311343193, 0.04524870216846466, 0.042636722326278687, 0.11206140369176865, -0.10703498125076294, 0.01894192211329937, 0.02102338708937168, 0.015665778890252113, 0.028368234634399414, -0.20243903994560242, -0.014300930313766003, 0.040765050798654556, -0.04896768555045128, -0.01203515287488699, -0.020295456051826477, 0.00276294257491827, 0.09726700186729431, -0.02014281041920185, -0.013521070592105389, 0.016450906172394753, -0.012285929173231125, -0.09785959124565125, 0.21327732503414154, -0.07401564717292786, -0.1393154412508011, -0.13385669887065887, 0.015849487856030464, -0.01531589962542057, -0.007976102642714977, 0.031828220933675766, -0.08619199693202972, -0.029555052518844604, -0.09507950395345688, -0.00015988407540135086, -0.034749891608953476, 0.017263473942875862, 0.010951412841677666, -0.005193289369344711, 0.09095882624387741, -0.09846039861440659, 0.007742667570710182, -0.0028685382567346096, -0.03562391549348831, 0.05373062193393707, 0.02564264088869095, 0.07774359732866287, 0.12877053022384644, 0.000933137140236795, -0.0027206246741116047, -0.01820615492761135, 0.1700851172208786, -0.07972845435142517, -0.017893977463245392, 0.1188436895608902, -0.015353376045823097, 0.05734476447105408, 0.09602402150630951, 0.04465256631374359, -0.08130361139774323, 0.035874269902706146, 0.04456299915909767, -0.01741213910281658, -0.24014046788215637, -0.013197758235037327, -0.04160812869668007, -0.025278182700276375, 0.145509734749794, 0.03676043450832367, -0.03298744559288025, 0.08204006403684616, -0.040616948157548904, 0.007186159957200289, -0.004690542817115784, 0.09505606442689896, 0.054122745990753174, 0.04061390087008476, 0.10934276878833771, -0.01692076027393341, -0.05474098399281502, 0.02557704783976078, -0.00957062654197216, 0.2266402244567871, -0.019239243119955063, 0.16132886707782745, 0.040463440120220184, 0.14493948221206665, 0.007500224746763706, 0.08142603933811188, 0.02305355668067932, -0.02125972881913185, 0.022011438384652138, -0.05988278612494469, -0.04871785640716553, 0.038611460477113724, 0.009895416907966137, 0.053685177117586136, -0.12299085408449173, 0.019009288400411606, 0.01937275566160679, 0.2927206754684448, 0.050925206393003464, -0.33743399381637573, -0.09590678662061691, 0.015219619497656822, -0.06147409975528717, -0.08921221643686295, 0.008782749995589256, 0.13046199083328247, -0.1083100363612175, 0.04076237231492996, -0.108177050948143, 0.08495990931987762, -0.06492073088884354, -0.003506746841594577, 0.06151379272341728, 0.07695627957582474, -0.01691494882106781, 0.06753946095705032, -0.24765078723430634, 0.2800402045249939, -0.009358251467347145, 0.07986140251159668, -0.050955016165971756, 0.025053804740309715, 0.03639324754476547, -0.029665153473615646, 0.07045823335647583, -0.005687389988452196, -0.09737948328256607, -0.19528339803218842, -0.10128507763147354, 0.03569556400179863, 0.12427563965320587, -0.058828361332416534, 0.13602349162101746, -0.03604944795370102, 0.0073720430955290794, 0.06212800368666649, -0.09303407371044159, -0.13535740971565247, -0.10054873675107956, 0.03150146082043648, 0.03208528086543083, 0.047882191836833954, -0.11709049344062805, -0.11449534446001053, -0.042663633823394775, 0.15494923293590546, -0.06470133364200592, -0.050394605845212936, -0.14482691884040833, 0.0633145123720169, 0.15359677374362946, -0.06739650666713715, 0.03594150394201279, 0.008239415474236012, 0.15234221518039703, 0.031054211780428886, -0.012922720052301884, 0.0799308568239212, -0.08632301539182663, -0.22121351957321167, -0.04557819664478302, 0.1498667448759079, 0.04406492039561272, 0.04713589698076248, -0.00811198353767395, 0.01162707433104515, -0.024175308644771576, -0.07702375948429108, 0.060396622866392136, 0.02427705004811287, 0.032721567898988724, 0.051192160695791245, -0.06311626732349396, 0.024148374795913696, -0.06772248446941376, -0.040859419852495193, 0.13723810017108917, 0.3204648792743683, -0.052150461822748184, 0.0006890312652103603, 0.027094626799225807, -0.06555260717868805, -0.1315649151802063, 0.01809152215719223, 0.12180569767951965, 0.023751163855195045, 0.03215234354138374, -0.19521938264369965, 0.06152581423521042, 0.09140288084745407, -0.02338082157075405, 0.08109793812036514, -0.28715091943740845, -0.13152189552783966, 0.10188654810190201, 0.15205703675746918, 0.0026508618611842394, -0.16504009068012238, -0.056855861097574234, -0.010048494674265385, -0.07525187730789185, 0.10148102045059204, -0.02241622470319271, 0.12348797917366028, -0.018514828756451607, 0.05824681743979454, 0.035330772399902344, -0.06880667805671692, 0.16756239533424377, -0.03523188829421997, 0.07200338691473007, -0.018795888870954514, 0.034990888088941574, 0.047701068222522736, -0.0753806009888649, 0.022855665534734726, -0.08411231637001038, 0.0455201119184494, -0.1361072063446045, -0.03473970293998718, -0.07774227112531662, 0.03489946201443672, -0.05092055723071098, -0.03915530815720558, -0.00012297951616346836, 0.04532131180167198, 0.08019205927848816, 0.0025581212248653173, 0.13100579380989075, -0.02217167615890503, 0.14295926690101624, 0.0846996158361435, 0.11009250581264496, 0.014257832430303097, -0.05856245383620262, -0.039550602436065674, -0.009058895520865917, 0.04608170688152313, -0.10616566240787506, 0.04143628105521202, 0.1271659880876541, 0.035782672464847565, 0.14859040081501007, 0.0561017170548439, -0.0640990361571312, 0.009064273908734322, 0.05981213599443436, -0.10252760350704193, -0.11076215654611588, -0.022697526961565018, 0.07480943948030472, -0.1784818172454834, -0.0032824263907969, 0.11760550737380981, -0.06606131047010422, -0.011229893192648888, -0.008461114019155502, 0.02472645603120327, -0.013034890405833721, 0.19426581263542175, 0.03926347941160202, 0.07663396745920181, -0.06778346002101898, 0.08797430992126465, 0.03949011117219925, -0.1360609233379364, 0.04591786488890648, 0.05430617555975914, -0.04611017182469368, -0.026967985555529594, 0.049629006534814835, 0.11067333072423935, -0.005649589002132416, -0.044913873076438904, -0.12603458762168884, -0.11256633698940277, 0.06122264638543129, 0.08777856081724167, 0.03173816576600075, 0.02794795110821724, -0.01371944509446621, 0.029535464942455292, -0.12841325998306274, 0.11754646897315979, 0.09101279824972153, 0.09496205300092697, -0.14831951260566711, 0.1508052796125412, -0.006085787434130907, -0.0021594970021396875, -0.006499950308352709, 0.020906636491417885, -0.10175158828496933, 0.0003838474804069847, -0.09298834204673767, 0.01867212913930416, -0.055678535252809525, -0.003647899953648448, 0.013598051853477955, -0.05725730210542679, -0.06277346611022949, 0.005747925024479628, -0.10373909771442413, -0.05525960400700569, 0.014823894947767258, 0.07042127102613449, -0.09229522943496704, -0.0258860494941473, 0.04872816801071167, -0.11346448212862015, 0.0745752602815628, 0.04738625884056091, 0.019277187064290047, 0.024776551872491837, -0.12600891292095184, 0.04914962127804756, 0.02410060167312622, -0.010405877605080605, 0.003520002355799079, -0.15054266154766083, 0.006273633800446987, -0.013039014302194118, 0.009344813413918018, 0.006199652794748545, 0.043612007051706314, -0.1380995362997055, -0.02914389595389366, -0.02172836661338806, -0.038754407316446304, -0.06883009523153305, 0.03820078819990158, 0.02227592095732689, 0.016986709088087082, 0.18065378069877625, -0.09153429418802261, 0.042596034705638885, -0.2344827502965927, 0.007699972949922085, -0.015221689827740192, -0.08589351922273636, -0.037619128823280334, -0.02936384081840515, 0.07975896447896957, -0.06882591545581818, 0.09098803251981735, -0.04299096763134003, 0.019151750952005386, 0.02753213234245777, -0.09738482534885406, 0.0498037151992321, 0.03593948483467102, 0.2669641077518463, 0.03312455862760544, -0.033440034836530685, 0.05693145841360092, 0.0009194828453473747, 0.058837905526161194, 0.10473573952913284, 0.15200138092041016, 0.188407301902771, -0.029375165700912476, 0.09476368874311447, 0.032651741057634354, -0.08191924542188644, -0.10335778445005417, 0.09818553179502487, -0.030119098722934723, 0.10034479200839996, -0.011954461224377155, 0.22330228984355927, 0.12311773747205734, -0.17943815886974335, 0.017866874113678932, -0.026652444154024124, -0.07977089285850525, -0.09269604831933975, -0.0780022144317627, -0.07913077622652054, -0.13980768620967865, 0.011135057546198368, -0.12405958771705627, 0.018075168132781982, 0.06963111460208893, 0.02177444100379944, 0.005763449240475893, 0.16638870537281036, 0.06673870980739594, 0.0043214610777795315, 0.11157507449388504, 0.014266100712120533, -0.0013423244236037135, -0.051854681223630905, -0.1263914406299591, 0.044787921011447906, -0.01657617837190628, 0.05085892975330353, -0.04980757087469101, -0.043657392263412476, 0.06355071067810059, 0.004942154046148062, -0.12326095998287201, 0.017119748517870903, -0.009260803461074829, 0.058882951736450195, 0.041154030710458755, 0.026054607704281807, 0.001776806777343154, -0.031635839492082596, 0.23827701807022095, -0.06681004911661148, -0.023101765662431717, -0.12400390207767487, 0.20816107094287872, -0.004809131380170584, -0.03775157034397125, 0.028650255873799324, -0.07925068587064743, 0.029600493609905243, 0.1543155312538147, 0.09143247455358505, -0.020598508417606354, -0.008840040303766727, 0.00008664511551614851, -0.014030077494680882, -0.04069508984684944, 0.10612049698829651, 0.0874619409441948, -0.014673602767288685, -0.06957411766052246, -0.04133161902427673, -0.045922014862298965, -0.028898131102323532, -0.044285666197538376, 0.059907086193561554, 0.03339741751551628, 0.0005226372159086168, -0.040447965264320374, 0.09603279829025269, -0.04146372154355049, -0.09493667632341385, 0.014681056141853333, -0.14599071443080902, -0.17090244591236115, -0.03680263087153435, 0.0657692477107048, 0.005300553515553474, 0.05034269019961357, -0.019815433770418167, -0.007301448378711939, 0.08060689270496368, -0.01024949923157692, -0.022752851247787476, -0.06627291440963745, 0.055370353162288666, -0.07734858244657516, 0.22861163318157196, -0.027462512254714966, 0.01868801936507225, 0.12820638716220856, 0.019359463825821877, -0.1102670282125473, 0.05177268385887146, 0.06717005372047424, -0.097547248005867, 0.06756767630577087, 0.20176541805267334, -0.033090051263570786, 0.10815050452947617, 0.058178648352622986, -0.10468325763940811, 0.0013395348796620965, -0.08885696530342102, -0.043697383254766464, -0.06415393948554993, -0.008936046622693539, -0.03863267973065376, 0.14446641504764557, 0.22085067629814148, -0.06960435956716537, -0.020878273993730545, -0.05918016657233238, 0.03458529710769653, 0.0580771267414093, 0.10296961665153503, -0.02611352875828743, -0.27710482478141785, 0.013617281801998615, 0.040779076516628265, 0.017148487269878387, -0.2737976908683777, -0.08833388239145279, 0.017733341082930565, -0.06416366249322891, -0.07316935062408447, 0.11136313527822495, 0.07947742938995361, 0.0498933382332325, -0.04864731431007385, -0.08964138478040695, -0.048577990382909775, 0.18462564051151276, -0.1563979834318161, -0.06439283490180969 ]
null
null
transformers
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed]
{"library_name": "transformers", "tags": []}
text-generation
datalama/gpt2-imdb-pos-v2
[ "transformers", "safetensors", "gpt2", "text-generation", "arxiv:1910.09700", "autotrain_compatible", "endpoints_compatible", "text-generation-inference", "region:us" ]
2024-02-10T16:28:24+00:00
[ "1910.09700" ]
[]
TAGS #transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us
# Model Card for Model ID ## Model Details ### Model Description This is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated. - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ "TAGS\n#transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact" ]
[ 57, 6, 3, 82, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4 ]
[ "passage: TAGS\n#transformers #safetensors #gpt2 #text-generation #arxiv-1910.09700 #autotrain_compatible #endpoints_compatible #text-generation-inference #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\nThis is the model card of a transformers model that has been pushed on the Hub. This model card has been automatically generated.\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact" ]
[ -0.05622259899973869, 0.16002345085144043, -0.004987028427422047, 0.023115945979952812, 0.0962471067905426, 0.011845538392663002, 0.06785304099321365, 0.11496778577566147, -0.020396295934915543, 0.11142492294311523, 0.03292480856180191, 0.0972127765417099, 0.11474913358688354, 0.16215258836746216, 0.004439093638211489, -0.23455148935317993, 0.04782992601394653, -0.12695099413394928, -0.033447545021772385, 0.11785799264907837, 0.14491069316864014, -0.10402194410562515, 0.07766910642385483, -0.030544815585017204, -0.009361269883811474, -0.03290390968322754, -0.06365230679512024, -0.05152205005288124, 0.05037128925323486, 0.06932847946882248, 0.06591591984033585, 0.007509593386203051, 0.09122733771800995, -0.2655104100704193, 0.02280162274837494, 0.07630051672458649, -0.0015554219717159867, 0.07497020810842514, 0.048351652920246124, -0.08209776133298874, 0.0788840726017952, -0.05696587264537811, 0.14718368649482727, 0.08216129243373871, -0.08924587815999985, -0.1965435892343521, -0.08464295417070389, 0.10284840315580368, 0.18357418477535248, 0.05158785358071327, -0.024141347035765648, 0.10476154088973999, -0.08419200032949448, 0.008797040209174156, 0.06024181470274925, -0.06443428993225098, -0.05412506312131882, 0.06934051215648651, 0.07975570857524872, 0.07967228442430496, -0.13025140762329102, -0.014651902951300144, 0.011243549175560474, 0.007594773545861244, 0.08504551649093628, 0.022028017789125443, 0.14595499634742737, 0.04393624886870384, -0.13030564785003662, -0.044304780662059784, 0.09771761298179626, 0.04345165938138962, -0.053857799619436264, -0.2537047266960144, -0.024983759969472885, -0.03927002474665642, -0.03094942681491375, -0.038562554866075516, 0.04431856796145439, -0.011080716736614704, 0.08032315224409103, -0.01118796318769455, -0.08149448037147522, -0.041395120322704315, 0.06544242054224014, 0.062143467366695404, 0.026896316558122635, -0.01158317644149065, 0.00973866879940033, 0.1224486380815506, 0.10907839238643646, -0.12763150036334991, -0.05768941715359688, -0.06755511462688446, -0.08307720720767975, -0.04300352931022644, 0.03337155282497406, 0.044020529836416245, 0.04436098039150238, 0.2466370165348053, 0.01108562108129263, 0.05453123152256012, 0.045806169509887695, 0.010608446784317493, 0.06787561625242233, 0.11606968939304352, -0.062306761741638184, -0.09178462624549866, -0.029058339074254036, 0.09215214103460312, 0.006741520017385483, -0.042814407497644424, -0.060904473066329956, 0.06479041278362274, 0.012608112767338753, 0.12110785394906998, 0.08444269746541977, 0.0026690615341067314, -0.07305197417736053, -0.06963318586349487, 0.18848419189453125, -0.1598394364118576, 0.047875016927719116, 0.031182926148176193, -0.038971830159425735, -0.0014042917173355818, 0.008752269670367241, 0.02394084818661213, -0.020246321335434914, 0.08923295140266418, -0.05574449151754379, -0.03784004598855972, -0.11079790443181992, -0.03252100944519043, 0.030985163524746895, 0.0051483530551195145, -0.027043871581554413, -0.033837489783763885, -0.09040277451276779, -0.059588029980659485, 0.0922931432723999, -0.07471107691526413, -0.04984431713819504, -0.013726521283388138, -0.07691634446382523, 0.023329194635152817, 0.016799474135041237, 0.08357251435518265, -0.02157396264374256, 0.0384126678109169, -0.0560205839574337, 0.0631464347243309, 0.11269522458314896, 0.029363946989178658, -0.053069718182086945, 0.05750001594424248, -0.24315528571605682, 0.10326608270406723, -0.07320205867290497, 0.050549428910017014, -0.15059062838554382, -0.026000602170825005, 0.044471126049757004, 0.00805877335369587, -0.013138634152710438, 0.14088952541351318, -0.21621745824813843, -0.0323486253619194, 0.16741067171096802, -0.0939871072769165, -0.07602590322494507, 0.059108685702085495, -0.05233629792928696, 0.10869261622428894, 0.04351044446229935, -0.02232111617922783, 0.060673557221889496, -0.14475463330745697, -0.01067100279033184, -0.04139741137623787, -0.02402937039732933, 0.16397778689861298, 0.07567544281482697, -0.06286642700433731, 0.08052356541156769, 0.024165838956832886, -0.017831770703196526, -0.04484899342060089, -0.023361295461654663, -0.10819391161203384, 0.009856974706053734, -0.06032416597008705, 0.02424289658665657, -0.025761527940630913, -0.09367526322603226, -0.02868773601949215, -0.1802000105381012, -0.009223134256899357, 0.0881323292851448, -0.011722641065716743, -0.021903391927480698, -0.12039245665073395, 0.011948852799832821, 0.031212422996759415, 0.002984174294397235, -0.13029038906097412, -0.05838731303811073, 0.027675874531269073, -0.16422230005264282, 0.03272955119609833, -0.05597274377942085, 0.05056252330541611, 0.03445037454366684, -0.03187771514058113, -0.033117350190877914, 0.009550533257424831, 0.006354342680424452, -0.010578392073512077, -0.2502359449863434, -0.02440580166876316, -0.0219739843159914, 0.17386503517627716, -0.21793730556964874, 0.04213962331414223, 0.07686693966388702, 0.14929872751235962, 0.006240781396627426, -0.038500864058732986, 0.010139784775674343, -0.08222103863954544, -0.030560437589883804, -0.0643099993467331, -0.012082485482096672, -0.03717579320073128, -0.05608142167329788, 0.05165567249059677, -0.16133594512939453, -0.028727244585752487, 0.1057019829750061, 0.06860516220331192, -0.14001330733299255, -0.019125886261463165, -0.04171464592218399, -0.043496038764715195, -0.05877087265253067, -0.0552728995680809, 0.1185101792216301, 0.05596614256501198, 0.04696191847324371, -0.06956122815608978, -0.07775315642356873, 0.007865429855883121, -0.017090093344449997, -0.017978519201278687, 0.08920905739068985, 0.07311701774597168, -0.12023317068815231, 0.09247473627328873, 0.10194233059883118, 0.09365488588809967, 0.108615942299366, -0.017981963232159615, -0.08929306268692017, -0.04584396257996559, 0.02045595459640026, 0.013332244008779526, 0.14797501266002655, -0.01403066236525774, 0.056954506784677505, 0.03922648727893829, -0.01123172789812088, 0.012020308524370193, -0.09384570270776749, 0.027314940467476845, 0.034342724829912186, -0.020308034494519234, 0.03796098753809929, -0.04001156985759735, 0.019826533272862434, 0.08712323755025864, 0.04676510766148567, 0.04415108636021614, 0.011758276261389256, -0.04233846068382263, -0.10904491692781448, 0.173858180642128, -0.12615609169006348, -0.24583272635936737, -0.14115718007087708, 0.0015609683468937874, 0.04152948409318924, -0.009671499952673912, 0.003867273684591055, -0.07054664939641953, -0.11710625886917114, -0.0934595838189125, 0.018713686615228653, 0.04491026699542999, -0.07426843047142029, -0.0596279613673687, 0.059872306883335114, 0.03894329443573952, -0.14430272579193115, 0.022237464785575867, 0.047419775277376175, -0.09032250195741653, -0.006925572175532579, 0.08398029953241348, 0.06729988008737564, 0.17764869332313538, 0.009659109637141228, -0.021044570952653885, 0.03080335259437561, 0.21258224546909332, -0.14283664524555206, 0.11252175271511078, 0.14021345973014832, -0.09024007618427277, 0.08099348843097687, 0.1948828399181366, 0.039186809211969376, -0.10478170961141586, 0.03259138762950897, 0.02489176020026207, -0.028939135372638702, -0.25018003582954407, -0.0680207833647728, 0.002590036718174815, -0.04892077296972275, 0.07092583924531937, 0.0918794497847557, 0.09946957975625992, 0.015428726561367512, -0.09732488542795181, -0.08017807453870773, 0.0468163788318634, 0.10640767961740494, 0.0070237633772194386, -0.01532268337905407, 0.08905128389596939, -0.03260866180062294, 0.018378758803009987, 0.0954233929514885, 0.00412675691768527, 0.17459604144096375, 0.05586163327097893, 0.17767499387264252, 0.07751350849866867, 0.06634163856506348, 0.019167855381965637, 0.0069374511949718, 0.02067388966679573, 0.017508454620838165, -0.004214957356452942, -0.08522020280361176, -0.00457410141825676, 0.12029227614402771, 0.06321834027767181, 0.024303704500198364, 0.0137604009360075, -0.03941800817847252, 0.08438141644001007, 0.17332784831523895, 0.0020201504230499268, -0.18486954271793365, -0.07240456342697144, 0.07921045273542404, -0.0910051167011261, -0.10552998632192612, -0.03353073075413704, 0.03346012532711029, -0.1747758537530899, 0.02097497321665287, -0.017018353566527367, 0.10809773951768875, -0.13855572044849396, -0.018670624122023582, 0.06328251957893372, 0.07232730835676193, -0.0028869258239865303, 0.06308864802122116, -0.153975248336792, 0.1050168052315712, 0.016289174556732178, 0.06754438579082489, -0.09747608006000519, 0.10138221830129623, -0.006303760688751936, -0.007241528946906328, 0.13875643908977509, 0.010596190579235554, -0.05694379657506943, -0.08987913280725479, -0.10555228590965271, -0.008462639525532722, 0.12933635711669922, -0.15157614648342133, 0.0847775787115097, -0.028662750497460365, -0.043171048164367676, 0.0024383023846894503, -0.1199452206492424, -0.1302652359008789, -0.1875755488872528, 0.058235347270965576, -0.1366453617811203, 0.039557021111249924, -0.10582595318555832, -0.04340389743447304, -0.028466427698731422, 0.2041483372449875, -0.2317875325679779, -0.0682469978928566, -0.1541893482208252, -0.08429346233606339, 0.14446710050106049, -0.04730919376015663, 0.08914490789175034, -0.0013825427740812302, 0.19013537466526031, 0.024473950266838074, -0.02387205697596073, 0.10308998823165894, -0.09543927758932114, -0.19450686872005463, -0.08603953570127487, 0.15582145750522614, 0.13931062817573547, 0.03702725097537041, -0.004593946039676666, 0.029260434210300446, -0.020000332966446877, -0.12535293400287628, 0.025526588782668114, 0.1793687790632248, 0.07859015464782715, 0.023437971249222755, -0.025896867737174034, -0.10993997752666473, -0.06524094194173813, -0.0335373692214489, 0.02718053013086319, 0.18264614045619965, -0.07421271502971649, 0.1900695115327835, 0.13626199960708618, -0.05445687845349312, -0.1955246478319168, 0.018216576427221298, 0.040417760610580444, 0.010847307741641998, 0.03138056397438049, -0.2078717201948166, 0.09027513861656189, 0.0014845491386950016, -0.05172133818268776, 0.141556978225708, -0.174949511885643, -0.1512570083141327, 0.06491631269454956, 0.0364508256316185, -0.19348180294036865, -0.117862768471241, -0.08817066252231598, -0.046907443553209305, -0.17498233914375305, 0.10519181191921234, 0.016932250931859016, 0.009516867808997631, 0.03492651879787445, 0.02640140987932682, 0.011080757714807987, -0.03873949125409126, 0.19461296498775482, -0.02505207620561123, 0.029532426968216896, -0.08079101145267487, -0.06136554479598999, 0.0607450045645237, -0.05577658861875534, 0.07896649837493896, -0.020188091322779655, 0.012835816480219364, -0.1100873053073883, -0.0468425452709198, -0.027396185323596, 0.017321845516562462, -0.09195652604103088, -0.09473495930433273, -0.05146971344947815, 0.09373841434717178, 0.08845265954732895, -0.036603908985853195, -0.04043547809123993, -0.07348548620939255, 0.0325477197766304, 0.17183002829551697, 0.17659065127372742, 0.038550034165382385, -0.08084331452846527, -0.005880105309188366, -0.01188716571778059, 0.04436201974749565, -0.22519725561141968, 0.06208868324756622, 0.04557957127690315, 0.015879612416028976, 0.11362850666046143, -0.018783990293741226, -0.16298477351665497, -0.06594224274158478, 0.06143777072429657, -0.06664001196622849, -0.18599680066108704, 0.0032026967965066433, 0.058006007224321365, -0.1646854728460312, -0.037671029567718506, 0.042260222136974335, -0.0045668939128518105, -0.04300284758210182, 0.01627597212791443, 0.08071378618478775, 0.005054219625890255, 0.07112491130828857, 0.05733523517847061, 0.0842885971069336, -0.10417009145021439, 0.07519911974668503, 0.08007751405239105, -0.08229218423366547, 0.031453702598810196, 0.08910130709409714, -0.061817802488803864, -0.03069761022925377, 0.032593827694654465, 0.07753410935401917, 0.019773589447140694, -0.041717879474163055, 0.008655321784317493, -0.09745000302791595, 0.06339588761329651, 0.09504765272140503, 0.03549657016992569, 0.014742289669811726, 0.034356739372015, 0.04988397657871246, -0.07460241764783859, 0.11766603589057922, 0.022336218506097794, 0.01780087500810623, -0.044981084764003754, -0.05459042266011238, 0.032110098749399185, -0.022974027320742607, -0.010163158178329468, -0.03885438293218613, -0.07015778869390488, -0.018130742013454437, -0.15929651260375977, -0.014899281784892082, -0.04085385054349899, 0.007158880587667227, 0.02551902085542679, -0.03834335505962372, 0.007963370531797409, 0.012195355258882046, -0.07085035741329193, -0.061454467475414276, -0.022903166711330414, 0.09224231541156769, -0.16436699032783508, 0.025155464187264442, 0.08285263180732727, -0.12099926173686981, 0.09775067120790482, 0.021939631551504135, 0.0031351554207503796, 0.028338242322206497, -0.1542527824640274, 0.04096807911992073, -0.024365095421671867, 0.01272035762667656, 0.04409142583608627, -0.22033950686454773, 0.001463581225834787, -0.03818526118993759, -0.05954346805810928, -0.010227864608168602, -0.033079732209444046, -0.11291328817605972, 0.09883669763803482, 0.008058897219598293, -0.08219768106937408, -0.030809206888079643, 0.03451729565858841, 0.08243680745363235, -0.02608415111899376, 0.15152283012866974, 0.0016822130419313908, 0.07172226905822754, -0.17519205808639526, -0.021702464669942856, -0.011611736379563808, 0.02207101881504059, -0.014536668546497822, -0.015496513806283474, 0.042471300810575485, -0.02421419881284237, 0.19108575582504272, -0.026401294395327568, 0.038726791739463806, 0.06405707448720932, 0.01593620702624321, -0.014801506884396076, 0.10957890748977661, 0.05975057929754257, 0.02399693801999092, 0.022115202620625496, 0.007329683285206556, -0.039842452853918076, -0.014149460941553116, -0.19538825750350952, 0.06474217027425766, 0.1377464383840561, 0.08781574666500092, -0.01322576031088829, 0.07683692127466202, -0.10024392604827881, -0.12397097796201706, 0.11215250939130783, -0.06283260136842728, -0.007701667957007885, -0.06531554460525513, 0.13346771895885468, 0.14944057166576385, -0.18992236256599426, 0.06835456937551498, -0.06228158622980118, -0.05332518368959427, -0.11744599789381027, -0.1957325041294098, -0.055616896599531174, -0.056456826627254486, -0.014700124971568584, -0.048795297741889954, 0.07307228446006775, 0.05693497136235237, 0.012962869368493557, 0.003600025549530983, 0.0766802653670311, -0.015357231721282005, 0.0008028073934838176, 0.03077360987663269, 0.06600049883127213, 0.013312965631484985, -0.02929985709488392, 0.020537450909614563, -0.007275243755429983, 0.04005419462919235, 0.06378308683633804, 0.038119763135910034, -0.02801438421010971, 0.01591232419013977, -0.03770609200000763, -0.10940317064523697, 0.0409080907702446, -0.028551526367664337, -0.08112191408872604, 0.13721226155757904, 0.02428387477993965, 0.005870606284588575, -0.02180131897330284, 0.24582624435424805, -0.07231455296278, -0.09001907706260681, -0.1473579704761505, 0.10211005061864853, -0.04095151647925377, 0.06560079753398895, 0.04110138490796089, -0.10732010751962662, 0.013498948886990547, 0.12688814103603363, 0.15896959602832794, -0.044884394854307175, 0.020156091079115868, 0.03252736106514931, 0.003683826420456171, -0.04006262496113777, 0.05253688618540764, 0.0694650411605835, 0.14883354306221008, -0.04907030612230301, 0.08928520232439041, 0.005485867150127888, -0.10256236046552658, -0.03822692111134529, 0.11808354407548904, -0.017866896465420723, 0.018703164532780647, -0.057248231023550034, 0.11889533698558807, -0.059861693531274796, -0.23005777597427368, 0.06317704170942307, -0.0720362737774849, -0.14286935329437256, -0.021647587418556213, 0.07456772774457932, -0.017636949196457863, 0.02658887766301632, 0.07326807081699371, -0.07681973278522491, 0.19899281859397888, 0.038975972682237625, -0.05729197710752487, -0.05658522993326187, 0.0789351835846901, -0.114089734852314, 0.2792985737323761, 0.01164181251078844, 0.04984506592154503, 0.10365619510412216, -0.016686614602804184, -0.13768579065799713, 0.015234606340527534, 0.09244892746210098, -0.09004336595535278, 0.03869183734059334, 0.2132277488708496, -0.002569539239630103, 0.1152428612112999, 0.07714667171239853, -0.07265080511569977, 0.04592108353972435, -0.1130065843462944, -0.0718315914273262, -0.086885966360569, 0.09441597014665604, -0.07240451127290726, 0.14123490452766418, 0.12318195402622223, -0.053516924381256104, 0.010368985123932362, -0.031209774315357208, 0.04651070013642311, 0.007842876948416233, 0.10365527868270874, 0.010769560933113098, -0.18099099397659302, 0.022656621411442757, 0.018202748149633408, 0.10856854915618896, -0.17241089046001434, -0.09672945737838745, 0.04725200682878494, 0.001958663808181882, -0.059874359518289566, 0.1282012164592743, 0.057909298688173294, 0.04923510178923607, -0.043742597103118896, -0.017267800867557526, -0.009560109116137028, 0.13584671914577484, -0.10737434774637222, -0.0021453071385622025 ]
null
null
peft
# Model Card for Model ID <!-- Provide a quick summary of what the model is/does. --> ## Model Details ### Model Description <!-- Provide a longer summary of what this model is. --> - **Developed by:** [More Information Needed] - **Funded by [optional]:** [More Information Needed] - **Shared by [optional]:** [More Information Needed] - **Model type:** [More Information Needed] - **Language(s) (NLP):** [More Information Needed] - **License:** [More Information Needed] - **Finetuned from model [optional]:** [More Information Needed] ### Model Sources [optional] <!-- Provide the basic links for the model. --> - **Repository:** [More Information Needed] - **Paper [optional]:** [More Information Needed] - **Demo [optional]:** [More Information Needed] ## Uses <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. --> ### Direct Use <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. --> [More Information Needed] ### Downstream Use [optional] <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app --> [More Information Needed] ### Out-of-Scope Use <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. --> [More Information Needed] ## Bias, Risks, and Limitations <!-- This section is meant to convey both technical and sociotechnical limitations. --> [More Information Needed] ### Recommendations <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. --> Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. [More Information Needed] ## Training Details ### Training Data <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. --> [More Information Needed] ### Training Procedure <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. --> #### Preprocessing [optional] [More Information Needed] #### Training Hyperparameters - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision --> #### Speeds, Sizes, Times [optional] <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. --> [More Information Needed] ## Evaluation <!-- This section describes the evaluation protocols and provides the results. --> ### Testing Data, Factors & Metrics #### Testing Data <!-- This should link to a Dataset Card if possible. --> [More Information Needed] #### Factors <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. --> [More Information Needed] #### Metrics <!-- These are the evaluation metrics being used, ideally with a description of why. --> [More Information Needed] ### Results [More Information Needed] #### Summary ## Model Examination [optional] <!-- Relevant interpretability work for the model goes here --> [More Information Needed] ## Environmental Impact <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly --> Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). - **Hardware Type:** [More Information Needed] - **Hours used:** [More Information Needed] - **Cloud Provider:** [More Information Needed] - **Compute Region:** [More Information Needed] - **Carbon Emitted:** [More Information Needed] ## Technical Specifications [optional] ### Model Architecture and Objective [More Information Needed] ### Compute Infrastructure [More Information Needed] #### Hardware [More Information Needed] #### Software [More Information Needed] ## Citation [optional] <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. --> **BibTeX:** [More Information Needed] **APA:** [More Information Needed] ## Glossary [optional] <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. --> [More Information Needed] ## More Information [optional] [More Information Needed] ## Model Card Authors [optional] [More Information Needed] ## Model Card Contact [More Information Needed] ### Framework versions - PEFT 0.8.2
{"library_name": "peft", "base_model": "google/flan-t5-base"}
null
mohammedRiad/flanT5_summary_withPEFT
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:google/flan-t5-base", "region:us" ]
2024-02-10T16:30:19+00:00
[ "1910.09700" ]
[]
TAGS #peft #safetensors #arxiv-1910.09700 #base_model-google/flan-t5-base #region-us
# Model Card for Model ID ## Model Details ### Model Description - Developed by: - Funded by [optional]: - Shared by [optional]: - Model type: - Language(s) (NLP): - License: - Finetuned from model [optional]: ### Model Sources [optional] - Repository: - Paper [optional]: - Demo [optional]: ## Uses ### Direct Use ### Downstream Use [optional] ### Out-of-Scope Use ## Bias, Risks, and Limitations ### Recommendations Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. ## How to Get Started with the Model Use the code below to get started with the model. ## Training Details ### Training Data ### Training Procedure #### Preprocessing [optional] #### Training Hyperparameters - Training regime: #### Speeds, Sizes, Times [optional] ## Evaluation ### Testing Data, Factors & Metrics #### Testing Data #### Factors #### Metrics ### Results #### Summary ## Model Examination [optional] ## Environmental Impact Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). - Hardware Type: - Hours used: - Cloud Provider: - Compute Region: - Carbon Emitted: ## Technical Specifications [optional] ### Model Architecture and Objective ### Compute Infrastructure #### Hardware #### Software [optional] BibTeX: APA: ## Glossary [optional] ## More Information [optional] ## Model Card Authors [optional] ## Model Card Contact ### Framework versions - PEFT 0.8.2
[ "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ "TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-google/flan-t5-base #region-us \n", "# Model Card for Model ID", "## Model Details", "### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:", "### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:", "## Uses", "### Direct Use", "### Downstream Use [optional]", "### Out-of-Scope Use", "## Bias, Risks, and Limitations", "### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.", "## How to Get Started with the Model\n\nUse the code below to get started with the model.", "## Training Details", "### Training Data", "### Training Procedure", "#### Preprocessing [optional]", "#### Training Hyperparameters\n\n- Training regime:", "#### Speeds, Sizes, Times [optional]", "## Evaluation", "### Testing Data, Factors & Metrics", "#### Testing Data", "#### Factors", "#### Metrics", "### Results", "#### Summary", "## Model Examination [optional]", "## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:", "## Technical Specifications [optional]", "### Model Architecture and Objective", "### Compute Infrastructure", "#### Hardware", "#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:", "## Glossary [optional]", "## More Information [optional]", "## Model Card Authors [optional]", "## Model Card Contact", "### Framework versions\n\n- PEFT 0.8.2" ]
[ 35, 6, 3, 54, 28, 3, 4, 9, 9, 10, 42, 20, 3, 4, 5, 9, 11, 13, 3, 12, 5, 4, 5, 3, 4, 9, 53, 9, 8, 6, 3, 14, 8, 7, 9, 4, 11 ]
[ "passage: TAGS\n#peft #safetensors #arxiv-1910.09700 #base_model-google/flan-t5-base #region-us \n# Model Card for Model ID## Model Details### Model Description\n\n\n\n\n\n- Developed by: \n- Funded by [optional]: \n- Shared by [optional]: \n- Model type: \n- Language(s) (NLP): \n- License: \n- Finetuned from model [optional]:### Model Sources [optional]\n\n\n\n- Repository: \n- Paper [optional]: \n- Demo [optional]:## Uses### Direct Use### Downstream Use [optional]### Out-of-Scope Use## Bias, Risks, and Limitations### Recommendations\n\n\n\nUsers (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.## How to Get Started with the Model\n\nUse the code below to get started with the model.## Training Details### Training Data### Training Procedure#### Preprocessing [optional]#### Training Hyperparameters\n\n- Training regime:#### Speeds, Sizes, Times [optional]## Evaluation### Testing Data, Factors & Metrics#### Testing Data#### Factors#### Metrics### Results#### Summary## Model Examination [optional]## Environmental Impact\n\n\n\nCarbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019).\n\n- Hardware Type: \n- Hours used: \n- Cloud Provider: \n- Compute Region: \n- Carbon Emitted:## Technical Specifications [optional]### Model Architecture and Objective### Compute Infrastructure#### Hardware#### Software\n\n\n\n[optional]\n\n\n\nBibTeX:\n\n\n\nAPA:## Glossary [optional]## More Information [optional]## Model Card Authors [optional]## Model Card Contact### Framework versions\n\n- PEFT 0.8.2" ]
[ -0.10398757457733154, 0.19306932389736176, -0.0033632912673056126, 0.03610021620988846, 0.09311320632696152, 0.019476331770420074, 0.05290861800312996, 0.12298566848039627, -0.03602222725749016, 0.10467136651277542, 0.0671035572886467, 0.1004098504781723, 0.10048767179250717, 0.2074693739414215, 0.009997162036597729, -0.2014479637145996, 0.02573133073747158, -0.09622545540332794, -0.0071241888217628, 0.12386585026979446, 0.15069414675235748, -0.09764111787080765, 0.08002780377864838, -0.017271630465984344, -0.015033651143312454, -0.03610097989439964, -0.07323332130908966, -0.03479931131005287, 0.04908183217048645, 0.049722496420145035, 0.05843517184257507, -0.000349408044712618, 0.09007392078638077, -0.2684064209461212, 0.019250305369496346, 0.045029979199171066, -0.008975513279438019, 0.08506851643323898, 0.10048305243253708, -0.03828867897391319, 0.12682385742664337, -0.03063143417239189, 0.14293353259563446, 0.0806744396686554, -0.09505166858434677, -0.21642039716243744, -0.0698619857430458, 0.08384164422750473, 0.16989272832870483, 0.07861769944429398, -0.044019971042871475, 0.13517163693904877, -0.09949695318937302, 0.018555914983153343, 0.043889496475458145, -0.08139429986476898, -0.07413902878761292, 0.057606976479291916, 0.10544946044683456, 0.055013351142406464, -0.13776178658008575, -0.031396493315696716, 0.02265457808971405, 0.040423475205898285, 0.07999663054943085, 0.01876903884112835, 0.13736023008823395, 0.03289055451750755, -0.1497451364994049, -0.040654655545949936, 0.13147787749767303, 0.031259022653102875, -0.03831277787685394, -0.22220522165298462, 0.009499377571046352, -0.08122815936803818, -0.026542557403445244, -0.0536605641245842, 0.035453230142593384, 0.000888594426214695, 0.09080729633569717, -0.03307853266596794, -0.09369251877069473, -0.01224873773753643, 0.09523900598287582, 0.04748715087771416, 0.024626968428492546, -0.025424735620617867, 0.007595535833388567, 0.1236117035150528, 0.05585745722055435, -0.12520839273929596, -0.05818035081028938, -0.0705985426902771, -0.04574494808912277, -0.04782046750187874, 0.03335172310471535, 0.03211933746933937, 0.05891488865017891, 0.2493341863155365, -0.02777520939707756, 0.053476300090551376, 0.05444002151489258, 0.020337484776973724, 0.04311012476682663, 0.09294150024652481, -0.05767429620027542, -0.1478586494922638, -0.016169501468539238, 0.09754423797130585, -0.010740936733782291, -0.02004612423479557, -0.048919931054115295, 0.03739603981375694, 0.045433323830366135, 0.10513025522232056, 0.09321142733097076, -0.003855899441987276, -0.07472645491361618, -0.049603261053562164, 0.20726995170116425, -0.14897844195365906, 0.040160760283470154, 0.022012438625097275, -0.01860237494111061, -0.05991404131054878, 0.012439980171620846, 0.01891867071390152, -0.024707071483135223, 0.09779436141252518, -0.06326582282781601, -0.04072722792625427, -0.1158904880285263, -0.020339088514447212, 0.03830961138010025, 0.013207120820879936, -0.02641453593969345, -0.03457428142428398, -0.06226963922381401, -0.0937361791729927, 0.10622812807559967, -0.06592761725187302, -0.06389700621366501, -0.030588660389184952, -0.0913286805152893, 0.01638869009912014, 0.0278138630092144, 0.1125856339931488, -0.02452649362385273, 0.04458776116371155, -0.014801345765590668, 0.061253540217876434, 0.07981594651937485, 0.03710449859499931, -0.06990481913089752, 0.060578957200050354, -0.1968754082918167, 0.09018965065479279, -0.07808025181293488, 0.029092613607645035, -0.1582079529762268, -0.01616949774324894, 0.0017850722651928663, 0.021649524569511414, 0.03714015707373619, 0.15085303783416748, -0.19596314430236816, -0.026611434295773506, 0.15674647688865662, -0.10309193283319473, -0.12404075264930725, 0.04324125126004219, -0.054548200219869614, 0.165940061211586, 0.021346256136894226, -0.008518904447555542, 0.0900673195719719, -0.15375599265098572, -0.022671451792120934, -0.02745574340224266, -0.00590722868219018, 0.10473140329122543, 0.08222995698451996, -0.07799220830202103, 0.02598065324127674, 0.01503553707152605, -0.04239313304424286, -0.023128151893615723, -0.05380116403102875, -0.11212947964668274, 0.002157418755814433, -0.08745094388723373, 0.02798321284353733, -0.007657070178538561, -0.07577638328075409, -0.012167530134320259, -0.16210347414016724, -0.02790268510580063, 0.07784745842218399, 0.01733841374516487, -0.019457129761576653, -0.09156484156847, 0.03273027762770653, -0.02178138494491577, -0.024860840290784836, -0.15668775141239166, -0.03685503825545311, 0.019179755821824074, -0.13811063766479492, 0.014558743685483932, -0.11879240721464157, 0.06631960719823837, 0.013130697421729565, -0.07176747918128967, -0.030934125185012817, -0.014247726649045944, 0.009652133099734783, -0.05069107189774513, -0.238661527633667, -0.020840344950556755, -0.05565066263079643, 0.15030357241630554, -0.2288348376750946, 0.040168363600969315, 0.041045330464839935, 0.127155601978302, 0.0029547119047492743, -0.05966789647936821, 0.0270998477935791, -0.07128892093896866, -0.022442305460572243, -0.07130090892314911, -0.00374257261864841, -0.0022334882523864508, -0.041810136288404465, 0.01707688346505165, -0.11181420087814331, -0.0432133674621582, 0.10227961093187332, 0.06174086779356003, -0.1733810007572174, -0.021279936656355858, -0.04796779528260231, -0.06899895519018173, -0.09009250998497009, -0.059767719358205795, 0.09957176446914673, 0.048852913081645966, 0.03765098378062248, -0.07288586348295212, -0.06857925653457642, 0.009726325049996376, -0.019988376647233963, -0.025859879329800606, 0.11403679847717285, 0.06938245892524719, -0.11552893370389938, 0.09631543606519699, 0.06578238308429718, 0.02606707066297531, 0.07889636605978012, -0.02589472196996212, -0.1066935807466507, -0.027396131306886673, 0.04618246853351593, 0.012227280996739864, 0.16259635984897614, -0.07902441918849945, 0.051724307239055634, 0.043603938072919846, -0.036024462431669235, 0.05044412240386009, -0.09820573031902313, 0.008799439296126366, 0.004851394798606634, -0.015200699679553509, 0.017037242650985718, -0.020525647327303886, 0.010141503997147083, 0.08521610498428345, 0.0529206320643425, 0.03382062166929245, 0.03404080495238304, -0.027921603992581367, -0.1317424476146698, 0.1793212592601776, -0.09554199129343033, -0.2353745996952057, -0.1547483503818512, 0.054340314120054245, 0.05688054487109184, -0.018691174685955048, 0.027349106967449188, -0.052948277443647385, -0.10158686339855194, -0.07906006276607513, -0.004239518661051989, 0.02939290925860405, -0.061315082013607025, -0.07630869746208191, 0.05408061295747757, 0.04271324351429939, -0.11554655432701111, 0.03984318673610687, 0.06148746609687805, -0.00884910486638546, 0.007091672159731388, 0.05241211876273155, 0.08309168368577957, 0.17821915447711945, -0.012110534124076366, -0.0015353928320109844, 0.05095047876238823, 0.2799825072288513, -0.15958015620708466, 0.11373674869537354, 0.11819417029619217, -0.06454669684171677, 0.07713348418474197, 0.18975575268268585, 0.03214599937200546, -0.10059504210948944, 0.03916865214705467, 0.03466443717479706, -0.026516474783420563, -0.2731192111968994, -0.04573635384440422, -0.009621784090995789, -0.09927961230278015, 0.08116234093904495, 0.09077970683574677, 0.09253760427236557, 0.03941492363810539, -0.06325390934944153, -0.09260788559913635, 0.03864419832825661, 0.09859392046928406, -0.018006911501288414, 0.00411513214930892, 0.08164795488119125, -0.028266923502087593, 0.008344229310750961, 0.09489850699901581, -0.018978014588356018, 0.16652531921863556, 0.0501289889216423, 0.10428904742002487, 0.07827963680028915, 0.09490728378295898, -0.006410010624676943, 0.023844541981816292, 0.01873089000582695, 0.022925296798348427, 0.0153488889336586, -0.08371207863092422, 0.03605816513299942, 0.10989228636026382, 0.03984716907143593, 0.02460450679063797, 0.013235803693532944, -0.0453256219625473, 0.04836542159318924, 0.18773561716079712, 0.009852245450019836, -0.1987961232662201, -0.08099905401468277, 0.053513053804636, -0.07738979905843735, -0.1403307020664215, -0.017406610772013664, 0.029183782637119293, -0.16918706893920898, 0.013617520220577717, -0.04516177624464035, 0.10099923610687256, -0.06584320962429047, -0.03969161957502365, 0.10316816717386246, 0.07027251273393631, -0.027419328689575195, 0.061324767768383026, -0.2036619633436203, 0.1284273862838745, 0.023495778441429138, 0.07020555436611176, -0.08861549943685532, 0.09673841297626495, 0.0015811880584806204, -0.006537758745253086, 0.1678459346294403, 0.002130899578332901, -0.07034989446401596, -0.05383555591106415, -0.09140164405107498, -0.013228128664195538, 0.1022302433848381, -0.1264127939939499, 0.06335079669952393, -0.01746957376599312, -0.03126668930053711, 0.006623926572501659, -0.07505890727043152, -0.13103319704532623, -0.17154714465141296, 0.05475964769721031, -0.10178133845329285, 0.03910680487751961, -0.09218303114175797, -0.06815411150455475, 0.004332813899964094, 0.1791655421257019, -0.17598746716976166, -0.09293381124734879, -0.1455414593219757, -0.08448480814695358, 0.16624417901039124, -0.036686088889837265, 0.086808942258358, 0.0021220718044787645, 0.16280865669250488, 0.013998540118336678, -0.000036381930840434507, 0.09926211088895798, -0.08773013949394226, -0.1966652274131775, -0.0585925467312336, 0.16685602068901062, 0.1420287936925888, 0.03928301855921745, -0.010226953774690628, 0.021435536444187164, -0.052637334913015366, -0.11157345026731491, 0.025610147044062614, 0.1282959282398224, 0.0841890424489975, -0.015405094251036644, -0.04002420976758003, -0.0957595705986023, -0.06662078201770782, -0.060585472732782364, 0.002853809855878353, 0.1857086718082428, -0.07526220381259918, 0.16260108351707458, 0.12022986263036728, -0.055326491594314575, -0.20451919734477997, 0.0500313825905323, 0.05739656090736389, 0.014329208992421627, 0.03546130657196045, -0.19687354564666748, 0.08637278527021408, -0.00028722695424221456, -0.07235647737979889, 0.16129066050052643, -0.16838403046131134, -0.14546513557434082, 0.09971419721841812, 0.03647508844733238, -0.23318703472614288, -0.13943128287792206, -0.09920816123485565, -0.017167937010526657, -0.11065550148487091, 0.06361348927021027, 0.001279310556128621, 0.01188784558326006, 0.0319657064974308, 0.01916741207242012, 0.029063472524285316, -0.04861828312277794, 0.20192812383174896, -0.02517242170870304, 0.009941681288182735, -0.050953514873981476, -0.09520260989665985, 0.03329018875956535, -0.049107227474451065, 0.10260879993438721, 0.0032580113038420677, 0.027790099382400513, -0.14269192516803741, -0.04454074427485466, -0.061776500195264816, 0.030475683510303497, -0.09861604869365692, -0.09038689732551575, -0.0458577424287796, 0.09854882210493088, 0.0978776067495346, -0.03187220171093941, 0.005129173863679171, -0.087631955742836, 0.07232155650854111, 0.2074398249387741, 0.18870997428894043, 0.07178229093551636, -0.061767544597387314, 0.021546823903918266, -0.03503970801830292, 0.0422513484954834, -0.22384749352931976, 0.04358934238553047, 0.05623139441013336, 0.02330639772117138, 0.08675912767648697, -0.009423524141311646, -0.15312659740447998, -0.07813537120819092, 0.08066268265247345, -0.047763094305992126, -0.16292329132556915, -0.027938414365053177, 0.03818659111857414, -0.2116163969039917, -0.048734840005636215, 0.015669941902160645, -0.02420305274426937, -0.040628574788570404, 0.024463245645165443, 0.07692307233810425, -0.019261030480265617, 0.10907949507236481, 0.0942755714058876, 0.09232617169618607, -0.09767459332942963, 0.07984596490859985, 0.0764341652393341, -0.047639865428209305, 0.02199367992579937, 0.11105269938707352, -0.05010578781366348, -0.0378328338265419, 0.09242516756057739, 0.08582907170057297, 0.02475828491151333, -0.05047622695565224, 0.015259156934916973, -0.05480033531785011, 0.06436607241630554, 0.12304303795099258, 0.02880598045885563, -0.004347979556769133, 0.058471813797950745, 0.03679032623767853, -0.0998108983039856, 0.11007150262594223, 0.06107207015156746, 0.022412311285734177, -0.04300978407263756, -0.025238243862986565, -0.01070684939622879, -0.009823714382946491, -0.019361721351742744, -0.005245421547442675, -0.09161651134490967, -0.008388632908463478, -0.10141261667013168, 0.03136153891682625, -0.07294134795665741, 0.01064629852771759, 0.028069494292140007, -0.05329771339893341, 0.008919995278120041, 0.00422483729198575, -0.07525654882192612, -0.050388675183057785, -0.012863846495747566, 0.08513447642326355, -0.12978507578372955, 0.03165426850318909, 0.07513056695461273, -0.10554037243127823, 0.0725056529045105, 0.001452068448998034, 0.006922660395503044, 0.01367163099348545, -0.1668684184551239, 0.057662345468997955, -0.02492786943912506, -0.013946481049060822, 0.014644047245383263, -0.2118615210056305, -0.011371584609150887, -0.04997337609529495, -0.05082135275006294, 0.013478727079927921, -0.027871666476130486, -0.12505188584327698, 0.10067178308963776, -0.004141174256801605, -0.0723273977637291, -0.016947394236922264, 0.0375068336725235, 0.09459720551967621, -0.023050690069794655, 0.12639659643173218, -0.028641996905207634, 0.07365536689758301, -0.17269618809223175, -0.0014769798144698143, -0.015675095841288567, 0.04053208604454994, -0.021747881546616554, -0.023975398391485214, 0.05700862780213356, -0.017800498753786087, 0.1788519024848938, -0.02125825732946396, 0.07307326793670654, 0.0551549568772316, 0.004134591203182936, 0.007016617339104414, 0.08551390469074249, 0.06321733444929123, -0.007615706417709589, -0.004623141139745712, 0.03993604704737663, -0.007357491180300713, -0.04575929045677185, -0.15855956077575684, 0.07038716971874237, 0.16100400686264038, 0.0457167811691761, 0.01973678544163704, 0.03137218579649925, -0.12055658549070358, -0.06745567172765732, 0.133295938372612, -0.006576597224920988, -0.03588211163878441, -0.07897653430700302, 0.18301799893379211, 0.13021139800548553, -0.1987198442220688, 0.07910817116498947, -0.06275517493486404, -0.05461886152625084, -0.12678936123847961, -0.1548735797405243, -0.06364554911851883, -0.0427422858774662, -0.02090756595134735, -0.061939842998981476, 0.051052600145339966, 0.051063172519207, 0.004142457153648138, -0.018934201449155807, 0.10347525775432587, 0.011050577275454998, -0.023741617798805237, 0.04846679046750069, 0.06263193488121033, 0.03206593915820122, -0.10124944150447845, 0.009767508134245872, -0.003167880931869149, 0.01839548908174038, 0.058993469923734665, 0.015632256865501404, -0.051406778395175934, 0.01211627945303917, -0.0161177646368742, -0.11322937905788422, 0.04360125586390495, -0.01821296289563179, -0.034581854939460754, 0.1380423605442047, 0.029063664376735687, 0.008010152727365494, -0.020634034648537636, 0.24100497364997864, -0.07432138919830322, -0.07833113521337509, -0.15428484976291656, 0.0600813664495945, -0.07019438594579697, 0.03150478005409241, 0.03358165919780731, -0.11591728776693344, 0.021276040002703667, 0.16450534760951996, 0.131715327501297, -0.0108525101095438, 0.011122293770313263, 0.05270205810666084, 0.0022694854997098446, -0.02992645651102066, 0.01490948535501957, 0.0539422407746315, 0.1351376175880432, -0.07271390408277512, 0.06536019593477249, -0.012350155971944332, -0.07296981662511826, -0.01739068143069744, 0.11009948700666428, 0.0008403394604101777, 0.0014486458385363221, -0.07303190231323242, 0.14288486540317535, -0.08721441775560379, -0.23417286574840546, 0.057911455631256104, -0.06816388666629791, -0.14905822277069092, -0.04948572814464569, 0.011390609666705132, -0.013741533271968365, 0.014513183385133743, 0.07532953470945358, -0.0468028225004673, 0.16993410885334015, 0.04449557885527611, -0.054205261170864105, -0.08118138462305069, 0.06101912260055542, -0.1261235922574997, 0.28507930040359497, 0.019505618140101433, 0.050904445350170135, 0.1038186252117157, -0.017813852056860924, -0.139632448554039, 0.011928204447031021, 0.106891930103302, -0.0686110183596611, 0.061951134353876114, 0.1714162528514862, -0.00013319606659933925, 0.12708258628845215, 0.05284430831670761, -0.05552060902118683, 0.040144626051187515, -0.0841156393289566, -0.052057117223739624, -0.11079906672239304, 0.07979317754507065, -0.08206817507743835, 0.16072049736976624, 0.13283812999725342, -0.06667226552963257, -0.0040592895820736885, -0.022347591817378998, 0.08292268961668015, 0.00745665468275547, 0.11128955334424973, 0.0083696898072958, -0.18610377609729767, 0.03635425120592117, 0.011172392405569553, 0.10025734454393387, -0.2108573615550995, -0.06490889191627502, 0.05169389769434929, -0.02080415189266205, -0.07541045546531677, 0.11757446080446243, 0.04553303122520447, 0.03321337699890137, -0.0389341302216053, -0.04741378128528595, 0.003704914590343833, 0.15027885138988495, -0.11527704447507858, -0.008043887093663216 ]